var/home/core/zuul-output/0000755000175000017500000000000015144730002014521 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015144732457015505 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000215641015144732305020263 0ustar corecoreŴikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB )?K x6b}Wߟ/nm͊wqɻlOxN_~𒆷7̗8zTY\].f}嗷ovϷw_>on3cvX~egQBeH,nWb m/m}*L~AzHev_uαHJ2E$(Ͽ|/+k*z>p R⥑gF)49)(oՈ7_k0m^p9PneQn͂YEeeɹ ^ʙ|ʕ0MۂAraZR׏!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPF:c0Ys66q tH6#.`$vlLH}ޭA㑝V0>|J\Pg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\[.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €'} S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|]|[>ӸUKҳt17ä$ ֈm maUNvS_$qrMY QOΨN!㞊;4U^Z/ QB?q3En.اeI"X#gZ+Xk?povR]8~깮$b@n3xh!|t{: CºC{ 8Ѿm[ ~z/9آs;DPsif39HoN λC?; H^-¸oZ( +"@@%'0MtW#:7erԮoQ#% H!PK)~U,jxQV^pΣ@Klb5)%L%7׷v] gv6دϾDD}c6  %T%St{kJ_O{*Z8Y CEO+'HqZY PTUJ2dic3w ?YQgpa` Z_0΁?kMPc_Ԝ*΄Bs`kmJ?t 53@հ1hr}=5t;nt 9:I_|AאM'NO;uD,z҄R K&Nh c{A`?2ZҘ[a-0V&2D[d#L6l\Jk}8gf) afs'oIf'mf\>UxR ks J)'u4iLaNIc2qdNA&aLQVD R0*06V۽棬mpھ*V I{a 0Ҟҝ>Ϗ ,ȓw`Ȅ/2Zjǽ}W4D)3N*[kPF =trSE *b9ē7$ M_8.Ç"q ChCMAgSdL0#W+CUu"k"圀̲F9,,&h'ZJz4U\d +( 7EqڏuC+]CEF 8'9@OVvnNbm: X„RDXfיa }fqG*YƩ{P0K=( $hC=h2@M+ `@P4Re]1he}k|]eO,v^ȹ [=zX[tꆯI7c<ۃ'B쿫dIc*Qqk&60XdGY!D ' @{!b4ִ s Exb 5dKߤKߒ'&YILұ4q6y{&G`%$8Tt ȥ#5vGVO2Қ;m#NS8}d0Q?zLV3\LuOx:,|$;rVauNjk-ؘPꐤ`FD'JɻXC&{>.}y7Z,).Y톯h7n%PAUË?/,z_jx܍>М>ӗom$rۇnu~Y݇̇TIwӜ'}׃nxuoỴRZ&Yzbm ]) %1(Y^9{q"4e?x+ [Vz;E|d1&ږ/0-Vb=SSO|k1A[|gbͧɇد;:X:@;afU=Sru CK >Y%LwM*t{zƝ$;ȾjHim @tBODɆj>0st\t@HTu( v e`H*1aK`3CmF1K>*Mk{_'֜dN${OT-n,'}6ȴ .#Sqη9]5zoX#ZVOy4%-Lq6dACYm*H@:FUф(vcD%F"i ' VVdmcOTKpwq.M?m12N[=tuw}opYG]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%Egl1$9  ֲQ$'dJVE%mT{z`R$77.N|biarNJ(Bň0ae3V#b,PY0TEu1L/]MTB4$`H6NI\nbǛ*AyA\(u|@ [h-,j7gDTÎ4oWJ$j!fc̖F4BJ2ᮚ苮p(r%Q 6<$(Ӣ(RvA A-^dX? I,($F{ձ7*Oy 6EK( EF #31J8mN .TTF9㕴/5~RxCe,&v3,JE- ZF5%Da,Gܠ*qI@qlG6s푻jÝ$ >8ȕ$eZ1j[h0SH,qf<"${/ksBK}xnwDb%M6:K<~̓9*u᛹Q{FЖt~6S#G1(zr6<ߜ!?U\(0EmG4 4c~J~]ps/9܎ms4gZY-07`-Id,9õ԰t+-b[uemNi_󈛥^g+!SKq<>78NBx;c4<ニ)H .Pd^cR^p_G+E--ۥ_F]a|v@|3p%kzh|k*BBRib\J3Yn|뇱[FfP%M:<`pz?]6laz5`ZQs{>3ư_o%oU׆]YLz_s߭AF'is^_&uUm$[[5HI4QCZ5!N&D[uiXk&2Bg&Ս7_/6v_cd쿽d@eU XyX2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcB j>Pm[E[^oHmmU̸nG pHKZ{{Qo}i¿Xc\]e1e,5`te.5Hhao<[50wMUF􀍠PV?Yg"ź)\3mf|ܔMUiU|Ym! #'ukMmQ9Blm]TO1ba.XW x6ܠ9[v35H;-]Um4mMrW-k#~fؤϋu_j*^Wj^qM `-Pk.@%=X#|ۡb1lKcj$׋bKv[~"N jS4HOkeF3LPyi︅iWk! cAnxu6<7cp?WN $?X3l(?  'Z! ,Z.maO_Bk/m~ޖ(<qRfR"Au\PmLZ"twpuJ` mvf+T!6Ѓjw1ncuwo':o gSPC=]U҅yY9 &K<-na'Xk,P4+`Þ/lX/bjFO.= w ?>ȑ3n߿z,t s5Z/ Clo-` z?a~b mzkC zFȏ>1k*Dls6vP9hS  ehC.3 @6ijvUuBY hBnb[ Fr#D7ćlA!:X lYE>#0JvʈɌ|\u,'Y˲.,;oOwoj-25Hݻ7 li0bSlbw=IsxhRbd+I]Y]JP}@.供SЃ??w w@KvKts[TSa /ZaDžPAEư07>~w3n:U/.P珀Yaٳ5Ʈ]խ4 ~fh.8C>n@T%W?%TbzK-6cb:XeGL`'žeVVޖ~;BLv[n|viPjbMeO?!hEfޮ])4 ?KN1o<]0Bg9lldXuT ʑ!Iu2ʌnB5*<^I^~G;Ja߄bHȌsK+D"̽E/"Icƀsu0,gy(&TI{ U܋N5 l͖h"褁lm *#n/Q!m b0X3i)\IN˭% Y&cKoG w 9pM^WϋQf7s#bd+SDL ,FZ<1Kx&C!{P|Ռr,* ] O;*X]Eg,5,ouZm8pnglVj!p2֬uT[QyB402|2d5K: `Bcz|Rxxl3{c` 1nhJzQHv?hbºܞz=73qSO0}Dc D]ͺjgw07'㤸z YJ\Hb9Ɖ„2Hi{(2HFE?*w*hy4ޙM^٫wF(p]EwQzr*! 5F XrO7E[!gJ^.a&HߣaaQÝ$_vyz4}0!yܒ栒޹a% Ŋ X!cJ!A\ ?E\R1 q/rJjd A4y4c+bQ̘TT!kw/nb͵FcRG0xeO sw5TV12R7<OG5cjShGg/5TbW > ]~Wޠ9dNiee$V[\[Qp-&u~a+3~;xUFFW>'ǣC~방u)т48ZdH;j a]`bGԹ#qiP(yڤ~dO@wA[Vz/$NW\F?H4kX6)F*1*(eJAaݡ krqB}q^fn 8y7P  GRޠkQn>eqQntq"Occ°NRjg#qSn02DŔw:ؽ 5l)Fa/TTmCԤ{"9b{ywSXE*m#3U ùRIvޏrJ`k|wJKH:O*OKy`( ݢe*{ ua ȻݔhvOkU~OǠI/aǕ-JMX _.6KsjA Qsmd  O#F.Uf28ZAgy>y,d$C?v01q5e.Um>]RLa&r?+@6k&#l)I5_> ` D s5npo}/ؙq #a2V?X~.4O/'|/_|&q̑0dd4>vk 60D _o~[Sw3ckpkpLNa ^j 5*<&}kˢmqvۗj=<Tr=[ a^؃ È(<^=xZb [_tܡ&yЋ{ Sym^?̑sU~' Ԓ f\itu)b>5X -$s?{WƱPz;| \;_D[T/BI GH8@"t*"9E[/Y5d{zrBܖ6Hlc "mKv~[uLU4lZ;xEN'oI㤛rP*jC# 6@dmHg1$ʇȠh#CBΤ{sTQ{%w)7@y1K^ ].Y$46[B-3%OONw8d`Q4d$x0t8@t]y1T\YAidtxBG:pɨyeNg4n]M؞ e}Wn6׳i~'ہZ*FU{fXڃP'Hl4 ,ŸqMHDCYZz Qnz܁$Jp04ȴIL΃.0FiO-qy)i_TA|S2G4miBȨHM(2hys|F 94 DNlϒòκ-q|xC ,gKDzHR%t+E/wd#礱ºȄWEz o\JξB.wLKZ39(M +(PWՇfR6#ю3Ȋt ݪbh]MTw䀩S]'qf&)-_G;"1qz퇛0,#yiq$ՁɄ)KٮޓJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?,ڼ"ϴq*(A7? /W= #^ub"6q f+=^OI@߱^F[n4A#bYѤwd)J^Z{*ǥzw73LuaVad=$6)iI gC~.1%YmҪ+2gSt!8iIۛ*JgE7LGoş\bC}O i ycK1YhO6 /g:KT sPv6l+uN|!"VS^΄t*3b\N7dYܞLcn3rnNd8"is"1- ޑܧd[]~:'#;N(NknfV('I rcj2J1G<5 Nj̒Qh]ꍾZBn&Un' CyUM0nCj.&Oڣg\q0^Ϻ%4i" ZZG>Xr'XKc$2iσֹH<6N8HSg>uMik{Fm(W F@@{W+ߑ?X2hS4-=^YgpUHެbZ!y!ul@ڼ63" ۩:6=TZõ$E,ϓRV|G&$rr;J TtIHFE=RȬ]P pLm|?$%>Eü%mWO[>Xmw,*9.[G n >X8Ī;xW%dT:`ٓ~:QO,}j6j!yڦʲT:Pqҋh] H+&=>g| Z;D8ܶb:! Å{2:+au 6:!fF+0#+̬NY"!6a7#񕪰%:r|o5Znڧs?si/W qEU馥˟^_޶oڷOj'?nc]Rn\t3^邳塨Lɏ"8k8M~?M}OAH$77f|lgn I;.K*!<+"eK5c&`X:#;@B@[(K44sBFu M.MNWLlY]K᜴=/ VމYlϿ4i36$>m|_>9|dUA"{!$jKx E$K3hN(tÊ-#v#O N, 9g80Ǭ&VdӞ5W1!1KYd`,-*&>F~⯰&jb.~cNk BL_OG]Bv.A|'qT(Ol.' 4IE|@Iі)<-p JkQm1 `qacܗVc?)cl*&<}P媠E{-sVU>߇GUt\+n3X]Byoz)li$2cPs6D>TE-n# rve{椱I |p)U݋7yJw&PzDgi xs  xh\L r Ѥo Zt(I >|$>tnMdэoV#ہll/ؽnA(ȱbAj>C9O n6HNe">0]8@*0)QsUN8t^N+mXU q2EDö0^R) hCt{d}ܜFnԴ.2O⠪R/r| w,?VMqܙ7'qpUۚ5Tnj ۝jlN$q:w$U>tL)NC*<` `)ĉJآS2 z] +RlWDžuF7LFֆM~ar*EtIbW>jqour?qzJJaQ#-n`/$fhnqgTĔO5 ꐌSYXzv9[ezksA`<dkON৯s|&*pNaJه5B5H:W2% `6MRR'xZtfC$1aH_dx$1'/v^ZZ4`9);q`F"d1v>ժbLGd~MP%m x52LMF9 E"A,S Vo}\"X.2< 5FB΢u.`aJ#Tk’"D#cuCXȉ4 ՖK(KP|dZ1&8{9rLnMRф%V Ng2K|`ot.GSGd oE'!B'Nb1{8LW^9KbN;sö!`0ݘ/l+1L#B8U֕&*?V6N{I_Pĝ"] rT [eTr؟˰ ]\ h! v˱>5S1px fnk}sRmA>d2UAkؖvlX܇Bz1U_#Xӫ+al H d\k/I,k,ρ|`zR/$@8VU^rcG"E7\qtS:ڝUyy >Vc11*?xYa8U`Jw/AcL~|;yj8TR#s"Q.ϊ/Yrx+u6*27fǪC%+A~*Zآ'ѭnۡ|< a1s\ T5҃F<[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ'`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;LҼé1ivGgJ+u7Τ!ljK1SpHR>:YF2cU(77eGG\ m#Tvmە8[,)4\\=V~?C~>_) cxF;;Ds'n [&8NJP5H2Զj{RC>he:ա+e/.I0\lWoӊĭYcxN^SPiMrFI_"*l§,̀+ å} .[c&SX( $*׊Q v&P_񾅶X/)T/Rs9`RI/Hй qPq3TY'se$U>tLNÔO\c6:$[|tӒNc_:F*`Nv; +4 {e6J61Z!֨řoҒ"HKX 6„=z{Ҍ5+P1;ڇ69y*?Ԩ9m/r>8.fgW a}ev9ka-9`x 92$4_>9WvTIj`+C2"s%DƖ|2HCvQu0 M.1%]vRato Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY/FRĦ@x؂ǰA~б Xn#[r>$9;[?- й vRCxKVV+#lj@_RL;IQ8ŢΌXD@Z}\QbR9GuB/S5^fa;NwQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZt$d:z4 .}gRcƈ^ʮC^0l[hl"য*6 ny!HQ=GOf"8vAq&*țTOWse~ (5TX%/8vS:w}[ą qf2Lυi lm/+QD4t.P*2V J`\g2%tJ4vX[7g"z{1|\*& >Vv:V^S7{{u%[^g=pn]Y#&ߓTί_z7e&ӃCx;xLh+NOEp";SB/eWٹ`64F 2AhF{Ɩ;>87DǍ-~e;\26Lة:*mUAN=VޮL> jwB}ѹ .MVfz0Ïd0l?7- }|>TT%9d-9UK=&l&~g&i"L{vrQۻou}q}hn+.{pWEqws]]|/ǫ\}/J.MLmc ԗWrU}/Ǜ+sYn[ﯾeywyY]]¨Kpx c./mo;ߟRy*4݀wm&8֨OrmXMlk1=VzpO֠24hf 1hi D{q:v%̈#v^nBi~MefZF >:/?Ac 1Mo e9d ljB?K_z>p%'3JQK-͗R>KkΤOq,*I|0]Sj%|-Ԟ = Ʃ%>H&t;9`>$& n(Ȋ1{TT%41?md+F/׌N>68M`_IIv48ںQ{p#3PaNE2:BU,6{ eTFY:\D0~ PmRCUl |P &mPu U! D"/RnJr *J)B@qv3X(fuQVfk*0+nO/yPH> Ua*n"kQ"`-Ń2R,:qHi ~q }Un깘TRcQf2hVA< d͢_WWz 9> 9/q< fdz|I5L!. `0`P%3d@OIىGnXZyύ/C 0fύ/Op4|1Bq]׵ 0gel e`eٌmk0`&7<4yDdu~FA@uAcדX~@(A.8EAHYv=j mo-Ogȣ/ِnn|o5?r0g/M{t4 '(/q+聇 ѱ4l%gV*u=GsLݶ5wsxȣe=Mq4,a"8E>E?z$QY/p-3dDcFbTO8UIJaTMy P=!evj2p4(.5NmgTj)pl>}8Q?XW"mشq%IE !ՑnнeT,0zkQ`Y2W, !*"qW+h) թ:V+4P=wYt^Ȳgoxz;cB^.OPח*`N%úI[QŰȰ*LZYßLOۤ;Q"ʞЀe ۰v~k۰ʣ! u- bŵϣ4<k':gkBnmIs)}#oD%ދjAs._:`mtWк% MR2خc푝m݌܉ܤ"eat٧+/JC$^3:ςUtv",jd*=GڶY-%t? KV@hyÈ y݇W_4M H7;"Ql5}\@RU 7i*.,I*4b]"CKm-:>2tMsss/t Rڕ-(qIiК-xLIVtѫJ4+qx mWhDȢ]<f>8,(hbM"'`&x;fu4(LWC)!5q$+3"mQmuhsrxO }5a3"hOFj30Rߢ}܁*G` _`Dj*7JiT.wx25OE G w$8"k]>fq`ُs"wաQ2m%ouL"XFeVnh>nnYp\"u T%x_hF޲ٮH\= i,1cG@hg@]%hHd ȥ7h] UM%"J}ꖫRüԾ'DGeA'JhajIj\PRղQIaN2]8QܦB2c5QUj!z>-i.lډ8^X&Bgv"^?O7xy֗Z {&˃Yg[65zVePct@%B ](H߱B}/W)Tc:;@ oj kkIdճv޵{(rS UH˨}㝜b: Sm5ؔM+E4+Z*QE4=08'VPDqZr6+k([7fbEw%% KJ181.K-SM EVR#uh)b+dQDjlגk,1͊(P,lh( nkx4Uz2Wyڪ)UTI-b]4I*"nեV{*Ls/^7QkTUŦ:\QuEew N;pGut*iu a}>D6r 0Z(LMz /KNX.v\yAać(Wgw;ϖrY",W@Ah+aҞ08Vw_mg0ˍhu D&RjuID=!,Ou3iʲ"Y1i^|,vR4CfHilՉZM4Rb ltO65FXu ͱK (!櫚GbVǛbqbTvMzh,SSJ 6dE ?ֽ)n4h$y+lܝn%1ŔU,KNJ h:*B? ~) v>>RFVua~.'kK{kʅ0iY ͥi / ca3òYi|7Pprg;\^@cWGv4Ķ5?L\kU[it7+[fnTe-IZ%ẃe 563@R_lcҲ4xFV8MEa;, =>Dw1a9i NFix\[6ndsoiݰ5a&O_5ccHhna4lhMBCo⌛Fnb@1!L &aXQσi1Co581Zs c.X k;MI|oquNMnPW[nP8?.(Sc-sƸe3Q3[{pa¨@n@ƈr۴2-@-se,`qd: f#M5Fp1kJ:c8p0m783~AqFH?9.CiqgO93a8gyp{qb0䫅 8q;;Yy~ɜq|y! $]/% Fo u4:2[9apY6N<>1i:qG9iFs=sJnEB@]e uV:@qt;採||H }380?S|~hdtMƒY{TgF=ȷI<}-)={]$ծe(C$@kK6QAS>nXf}V;Oo@ٰ_ctTyс?Oѹ뗇G<* 0`Î"a.axmݍ3N"% :b Àvޙ OF{ W,.q]P5W;  LOeTLET"1x=7>1Om8Q|6xsp8N,m)A]@ @\&(ֵ>w"n9Xۍ0%Q$68ú P R[u"# #QID{xb>r! ݎ9QJ evb'kPlCQ-"'JP bVPЍ(fg)lȲSf-]j"MҖ [6WΪ~^VY.Q t JhF'}eS(qG4 pbI\䔪4r q}mW Dm g"Ty ˹Jysĸ\@,Fw jpKKا`߀`4X{CQ(<,i]ԸsP_s&?@Vk~̯13q=/1I,lV֣&P5򥔹Ȃr=7%6uV3$N @C)6ͩ#rZTsɺgU4^ceב86D =7ll_5_}"DɌAQ> h5B /KP~,ȸ,* (@ahU0RC E٠C;t;Q6΁?E%*-3 X}I:0=r6s q).XX]5& Ґޕ"`z!gyrqq tT^ǽ{qRڱy˸G4ܤH( + T%%Ke:?tȆЅ [ !̺Ϛgm]HwVHƇA/_OogѨ~%(ʈ8OΡQq2Of $"Bp"'iXm 5F wjȮɾΝ8̀ъSu ̮`!br@hjjQAզZ6}ZnyɾIPY]XPsm X<җkGkz^j[dr/ Z rtz1Ey*:\Grx[L0I1Vѹ8_G!;pdÜΗPE]#Ǔ騈whb*froN[I@*èN1<)CK˷~8 (<㌘; Ȋl^'YBF[`^f#&6{A(Aj'01a0O{`sAo qdK)6UmNuVRao`O9p =y?Ǖz'zN.*# ,ܧ{z$^vfB8Cnt6<~`27z 8\{[IM#Qĝ-E4?F4n#] xmzM;cv߱}2|Z|wZ~GRw ^%ށP{wBI*:GHBuW uw ݝPw?BG@Jz=PoBUBw'ߏP;@h;~$4؁ppB #4|$Btc3:0L)PLu|gF[ڇ̔-rZ\l4V=8LUQt6_9I tB*F$.q4Xf -@7R7po&|=eoϜ>9ﻴ8t_s0+fӬ|_`x50+*#WqyNCeGnW8 yrKn<8cxb26xvc`EB`V UMȁ~uzm~z:*$qPx;kV̟Zt1wNlvM20 f]zû04h| ㉒ Efu]߿Y1QυM H׬>\Sx1/ ^"{  ]-$yDI>=˒qOd?_h$~? c lsB=cN>`k3*BR  "7U 6N2&Xo;+-aLcҫh,@{.+wҽkNT<*<~'9lwlMAc`c@7^!R٬0C䶈躨=yAƐ *c0UFgA8*r#8v1emfx[(1bE&oO$9=+ tge$m5$H`) Z_xۢZ'csSH}_@_J`YD /kV&r{MA^!Ώs΍GCrit Osd̔mcz:vd0Kb}`Ѿ (> C Qn+Œ[|Cīzi*,6ହ[Yfd(+yra=-iVQ^޾uQXgx'PujDzlC򁵘_MվH{M,=P1z%%܂47֨B&[+pmXB}ՒV7q.F;t·{+p_ɒt(+5.Q p B;frv\qG=AH%8O] `u)AfRp BF=#8K;ʢ3pJ9.bԣeK #1&u׍EEy2~sJ%q~? n0l|}8d4ngu}tv{ªC]V5@78TjlgB(nTjiiyPwC+ӭDvPk4-Y8cMx\봡k,ma~94>q1nIQL* 1[VuI:ᵖ\(V{嗷^I= =Zë,gLTĴ"lsIQZuFyw;=@WMT, lxS7@DV  Wm1C>$#h,SeȻƎ_'Ue. vWE$G#R:#:rnȀmaz 1d|M?ԕ޶}=w^ZTt0&f.YA㗫<4nPjA9u[G|gҾ+~n~Vk8|5swvI ~iWj'_eDM|=~F|#UR Bfn>ir? }4^ebW4ICw?z2?^'/`iH | Yf^jhKc]6=Z-mC&۴ ghQ>IBÖn7}yN:*3%Cp.mc1+;(byG&of&"i)g<AX袑\ĘA[A V񬃶*Q~AfPܘEFbAi j^HM^jA(кP3_.{`5!Ֆ(-h$d4Dk\b h*Ht:]t)( Mf= rY#Zqs\zGHnv&l}<ܮ1m)L6xW!u0"[A4$IgQAR`#F´ B8Κ GCwx~Cn̷)` IFDw £kl0b iޭrɥI> Z0څi\z,5|)T:×5IrbIp2Lx~0Y+3q2 x93+<!S$]\7$X0U*ALrFj ARGl,p,jlS ]UgřCfĬ E4Ja)0]'ZnݺP/{Errt>'7<ɘZZuF)bA;Em?[iTg[MS=ģ({`+IU{pΥ˫&,] rQ%WH%'eaOv {ڙ _r],WARꏟیb/~rPEk)I&V=]VK\Ɛ m9xQ-#ELw?ư~L*8VU_ǨCQB'M$H:SKG'憘0:4sΘ^Xw 2FI -@0Kx yfJ áovЏC"YHSw $YZľr(^- vb"Yc5|&{_ .HnS|N 1)xI;Igx7wiL#){ݵ~]p<$1EĀԭ2YC5AG.3YJB)΃:v?vqp`>Uc 5y\ ҩ{t:*kJg那ɔ,-jfYH(ma7+Zgtգ8cR0]I[m1m1cceɳ6pXOZ@x sቡ Eҹxo]smM(mFitA#Kő )?.8`ֵD-*!P BCE}Iy-f$\6C_(B}7d8|ig׬C-14[ e\*bx+5h_**SKIg x]낣1',$,8K *Q&IR6i8j!tfEA 8kXUItF#"Ig8~#3)#gHVO]pL58գуJ͙)Vm^KӚH} ->>8dBC{=ئ{;|Vw:QѼ**#4G1 Qch, 2Eүؙ.-JFEe]IJ_Fn5p#paT88[51G111kEڳn$= z^Jx>>;8@4e3+QR6G, ~9h.nfݘKgg.u`}l8ftmD  y)'iS3\ŞRg.8ejB4ís ƽg]^;2_\;`GRׯs׃L{{KrQ3(be%<պB- u^G熄b[(So;v~YSv`ۋPeuӁ&D26pRyԲ%F׫M8fw(Sк)K R0A@+mXrP$TTK_u<Фwtg#tJRexVDۇ2 tE'>PI/t&ϴ7bCd0'I#|o` zx[SыQgIsѫʹȟIt>@.8%2e}e e[|U>ԜR` H:1^8xL%ZL(xOj) H"I|o]`Ʃ=^bx4A(v2,Gզjl:g/ iil$z)2AL j2O q> 1 Nh8Ĕ*+`]73Lt7Vt,jl_2hYW9C`JQ$~B_ /up%ɻ\NlpAp[+A*i˜5K,Q$s9EqQQz=?] Q(1p8et}[ULI `Ύ∝KJk,l5Fu8,ME"H;`"K^#O NX_8϶5Rb3kbnX1x¥] Fó5XAq .L T~SL.M.ot n4FIba%펏C(En#ۣH}rT`;qS+.B(k%'d*ݶǐ7W|lbȜ8 0dhJ( mVM-a a>W.U%ȕ7p7`j=e_ ̫S`űhCdpq5@siuBO7WwzY <ZsN IUdńJyuVEM;za}מ=8l.^.5%XT++~nip.^/0^Hɖ/rNSNX6#Tx肣4O%D^|>j%XO`6\Wr4 fhSxmb8'ԥ T Bx:4I ALwS &QUOsϔjw\p^ECspʷ4cƲ⥮/E5/J`">uqp}ҎӐ.ݮ)̼0ٻ6dW677 +ܬay|XBO?A(Îv~gFbLRHY"gkNW6*~R]Ѫt_'+4|GF[$u3ff>';QVKj)" *}\Ͱ m |TS[[/-.>rb$b=n IC o#[x[xs푦;tk][-܉M>#w5wwdŞ4=2Y3rϥܛnK [0;Ap ʚ'L9N`x۝m{Nř{"d:.(; SSՓ:.-g&_3 E_y׬iyQlhǻ;F h|ӷuǏ#td/ 9t32aF)FTSpnt*wDz:y \&s]s)TtJث{] C;gpUh5LR'V+O,Q\Kfm33Bp%)u \S=)BTz9 K xvznИviM@j6pI@aqe;MHar+ĒzM-` 8Ea(n [($kҪbYm7-g2Z .a ̰l6V x!݁yvc36h6yIڀ:]nIDH[/!c"uqm \QՉ*>' J#RT ڦF#G;I`,s r*u*65V4&y3Od^/ZɊF;(߿TdecCc}UV$pR6{VG۸q0c4n\ ; ^0U\{v]5br#ZM@7z;R"ܼ]$1膨57|D}*`0pهNi 5OⲤ3vyBjk+)+O+%^-KM]mzVyW.V-%޷ kckGݎ$G@82ƳG1RxtX;ɋ^xձ3t`9L92+ϨC"2X (A}3E}˖@ >t|{ sFJT[A9P~X{a!Xc6YG֒[D0NP!cXG1_J&]]g`1CKW^i]1$;z!05"De d_2-Jd}-g Sʎzb7VZG0=O E~R#-= 9g sG9u~=:Zfm{Z3[jq!;iWIukWQAZrx!RW6^e9Cc ^uuYq#bT倾{wqտʤZ3VfI 3aAcc#rlˌ^| X/ҏ,;: *+h_B_s\JF<Ԏ}nT/5@_OkG8\{B#"Dg0<_[w>/|%z/x +^oy~Y`ˏ_aTX><ʨ{-$h@(IQnU$i g0yI"oyDZ)K  sb=KWg5eA߶H[{m6z\Opm>z"qXEE?$5@SnS<5ࣸ5Ke< Baʒz ;Q):|qT ׅ7(=g?_mTӏi+ϛO7CEY xNǽ?Oh\1ʗo\{7cw&䄺i㘰!h82PVx\:f}t_,}7JdӆCQ Xn |jפl:ђ ׸8O5h ⠔%ǯ|PW28;beW* G8:=LAx( *hMj4(6xP%Lf`fh' P[7< B\ZI1 a-C;5iP#hٜ`KdjV9EjL6⶧H8䜬>|+ qq9M?ZuXjKb'ɛg|xQvWTW/di+` /E^0+Lh$**6s&f\ S18=2'jU;ؕ nXT gq a4i#ӫo-;o\|}?p"r0hOKꩇ ZsSwE888+Wg0M+oJ/P7 wJZ2 DWw |)i8|^E?kBP IJX6>SI;W/^ p!\3V7 ac=¹`(; qfKvfjyFR%3*3wFQ2FUǔ{ Ȇy.0Q"0&g6xm|4R5]"Ǐݚc*) {._W(J3qL"@F뻕3`maacK U!q#&4QǏњ$Rn+ld4Z"H* x|u8ovsESCyEp#eȝ1NrB 7}wTl*3jJV@_i15LޱZ3#Q\#kiɲ ߭]QsŘ 8̵d-Qs+֋5ZK06̖36 >mkmp`8j E(XhCRj*miTGI71+ dps:D&+HPeigf9W&Fj/Ė ۴jf7VLf4xLLzMi)BsG5/x{&c`G p5ƫ af $/Fe[G.0_SՆo7c~)(Z2NAEr;b4B#`P@|^IA!ढc% ܺ\r/X6|mN NR(2=@ӈ ^Rt!6bb85.0fLK[SPDn8sct C#9mBROaD p(ĕh)i'o?M tR{0`k*iNJi8S3Bv(Q86LJobHdt+FS$⹒ DIZhqse H#h$h4Fb|cKEYC%vBI#)2" |1O%B`XH&qy{\,̲cAZ(;h9&oA>[[x)$XkPp5أuٻ޶$W0Bw8`&L2X hM[Y }zP/%SfUuuW]*ei-+ 1-P*+`Z\[c8L`nE-DH% p$$1$)N%uK+ 6\GuPX"(%mcR'ebAl(FÑϺclb7SJrJKDHLyI"R}J<цFAD d4c!LĦ(N@IiJ"!bDŽh < ,YIbfjjd2"ϻΨ4 GGd*-_*ՖYM &Ý1 MUX),zt4BF.o=Bb i"mPhgz^Cg#Ͻca 5Αd[]NjɣX*r QQp'dI Dt1,6دi_ck?Nx}~Ɨ> I3?K8{!x0%w/Bq9t)$[|^=bXɋλyX1M\J]]=7MMn4o{wFODPK)f9f{6Qo.yze9Hi+ cMdgp'Er3 kY<^_`٬ 2_]äS׫\0 >jb5[Hśʈ@&ED6*LQӑj6DYZQ1rr= s SwŒG(SHn9^eO.%?4UҢ J擢,6> 'EVc9?]03>@@s)LOv|Sz|lE#rлf%k1WBqk =q~ٕlKFOj  aJpB9 3\H@H]Bъ6y90YVI5bvCg[UӑsY虌Ml=`ݯb I30[^i?`͡ӅdЀgP`P|=p av~|?#JrNl?;Y-?Jd7~zs^WZ {d+54/ݹnzU,łpnXpkuw'fb6ûِs V r|#'\ظGmNgu؟7i4wߵ̝n]> ٯcֆ.2]=RϦ[;x6sUcb3`)# UV=L ʰCEanՊ5k{w`!}䍈5*XC*-9}ΞЏO uWstzns[ WsD׷睼Al$ +<G$^;P+r3z4.:nи؝4d?͛,'){f%h|#B%%B H.nv{j%Lӣ;d1%|RzO ݨ1瘟i,eq&&41%RBǯ _R%BJdbd(vޏ⪑+=y # 9$ڑ6C 0TlL`/Ė gxM-t^SZ?y,*_R)Rӫ7КWk27I< BʕeEسZIEz_l;cx(2q>#gFTL`sӛQDŽj ͻ&fv!NFAv擛?KkzV_Yv\.z[y^M+¦g vM4sy!I1iDs V6qiBDrXԊ‚⑺Դ,$@੷$GN7/"?ˑ BԡN{0 N?/\Wًe4._G`-l~-ǴPq:2"7 3$}1ʮM7xc&?!&/hh7PBqu*v\ʏ:pWQWB{SX򛃼s7Cσs~ʜ479?Я@$ z_r A`ޛyMǛ?2geqK0-o^^c\;gn'\EK߿H^\@,HI-&Ge /yN_rd{ǣrל?6݂k~a{Hs!M]6PhZ-$3< y8yvSXJX/!c\tXhv+~?T´ &B#L[I1q}{-/[N^ 5qGä0L.>vx|~ӚfW6_x=EBw=us{Nοs&f;.x|}s}_w}]|OD.nV `.Y,VOԡ|Xקo@vhv;W.pR3iB;gCc$"KL͝ ,Ը0g&e]\ʀYC8F,JF"lpJ#+ő։@*ITDh' ex`]_>kk0ɟTwq/|e .V<XE6bo}Y <kMAp9K6>߇/b;/@ƹx좑uE+~,ׄwgײ|oV>۞pc<٢^ .=hȚ?`f+ ݖQJ] gy29k+vE[3v2?ggSW۪59&睵y7@q:=@>¿oj=0x (bƯ]V `-c,jjf/6>Wݭi%aɵy-hK1M?Ӓ3h=w^yRH~n i|y3̶́[?G"Vtd]oBHxjz},Yb K~_XrEoY'3KjW uSCb`P? ~KY@lvmNOhm;r!jv4 Ot6(*}E(Tð(d,81\+S'K!{84H',0<Ŕ&iZ+I  mjwvk힓V92[·.e~rTLlEUbBb;``tXY3 3:\"۹֜Oy`+g0-UaK!ߏ}d5ϊicsgŽEq#L7iBcE#^x)R%("bR؈ $)7)X'/u7?8])72A^e.+/WԞw/xra s_/;G]ʧwY@^Z|TȬaD FD,ɓ},9 T{R68(X!$aҦ( 0E4 X ¥;MpCL",#pSlSML\#r1Je fB5gfj{8_! n_bÖb_oeIVR~r8w4LꪙOsE0_6]ΆWJ8{モMoSXY?w<;HG|/b qD Ĥi!%@ >4H[Dsɯ!JrG|NR*LLWI+Yi)U46V<s o^Z_/~Gksw'NC%D:)ޟ!;~3a7a0MuJ#jYg[5 rۗZߍ֭va}'zP~\ɥNYe0.+ApD cHpђ*g%:)GZ|vۧhAk`rU|*2szSGMȤ% IW)]eXH=-q=!\9nC`c$`eN+bYR^TII=-q}!A^[~ ս06p+颐,1uDȉVQrJy 8KfC~惞2='RHNP\'|ԀiGT9z-yQ3Ɓ' ihrp~^fI3VqA>R:Ft= UC"2(+ʥ&zᗙ\R{5ä9,?q\Ca*@U 00 V^#-jDc)!&4u)AP)LdlQVĬ!y&bL6Ө*/z<sН``D[Ii~IA}! cA~g7qjCWY3pG|0}HV8˷Mۑ;'F"ߡ4%70KjI`^,2tCтGrG^"ѓpL%;GkE]vRC&" QlPD <#?Ѿ퇠{Bmp8tph((9C(e BHC:' i1iGWG5\g >p$$#aD NVqo9;{1 0F$lT:h}Еy{3 Z'y?em&VglP֜TUEI>?qUt. `Dx v b c@Ƒyw[׶^:s؂QIW60QCr 83jF^RY.Y:86rD={|1m[+p&/ #Iի !y`*U)DS\{ܛDR/z8 ^3;I9ÒQpjxZ2+`t9x"{7hP=A圍Q~>!C6BUJT*kK%8,=#xQ}!W&fsQ\8dQ3}m3pA Yi&M}w9/ -q ~-ß=FX%a3(w丈f Y" [>.J)o薳m}M_Kϳ;|ꫫl}o]ȳTU-k!ʟ#7zݵV*/{7@oƊ5$N 6+|JO~f%1ںrX%LTۉ&;Eb$7Dw)ijZZzj{ݳ*i5L:l;|gʍ:=5Y :ԍ~\^ #S:1)x}zr3Z5NyFQX֏<uÅk'W)8hL$E+đߵ.wכ*ûmhp95We1ܹD %NRA*0`tBht̕aRsN4My_p`iIڊ* _d-=41M`(h*6l)P5J8.lwΔH -!"(09vUL;i{nި`x 0[Ŵk/:9}(/xĜGU7&&@,5 r3G XyOY1e7$̗~U]<öapnr0\0ޅ9yM"Ɔ9c54og'O"Ir%[,lL@j='/GucĻq֌q89o8C͒=`=a-2T@ IʺW*)Y+?&'4N eyAĩmɾN(:N߁C;a}qK`>;#1A!cuQ8M]j7  gq;mmzwFA] qNGsSӟ5>ю{{6|3HHEN bT9%7ۧP}aoA*0D)/,GXіx;ge kͳ9h>" SBm;_6?j2Gl VJU3P1(XfYC]r\Ą7Wp) (#w|{N@h3%t%;:2BX/ RHI\zH&!̣W,]LvyOtՈc>OoX@U&1B1jPmHU)4D*L3t 7ٞlOy_p!7fŘ2&|[hA4Ga}Dpb'\Y8Vuј=7y_t}z.o>C-#N3ʹ˜|hejQHS4[Hx\Ϸ0 C/]P׾˙ü? g '58OPEODz9yӕi^@Ew,G`w ;bM;n"mT![/gC47T$U"? y_pA:`p#ޚZVgXS6npܺU:cA1#YaFù_C׋޳ph $o6|3\zw‹zRP1 3\;̵A2&|b<Ezp#jVN. aU*6|[i}yƼ9!c•hݜ1`me%Ww! 4pp^h<[g(7h}4jԌMjTD唇WNfV @ړջ3#1U8*Z!cua|Iuԕ2&|4rPN@=\r}- QuȘpx0#/uigwO˳PZ+љd%3/e1;l?Ѳ퇠b;[?8uM6.3Er'v6t|Bb!NR;V3v뮯s ٕpnʁ;d^D/rU30:dL猍uPRDTه 0]Ì>cHB T}=C= qH_a(7x*y)q(:Ms8Ka H⑄5NyQcް*32 HH NVq]{w*kT"1cqJvc3-p%1a2=B#q B+Zxt CN!;ҟ9>ԣٮ >y`0'+dX!nǿe04CťI1g_wg ,VzԒqs_/ D'aC9^Yt,ԛ ;xxov CܥȣO1p|,IQ_*)yt3V^Br6?0$ϠTC'?Ξs/TB^. _SCCH!Ʈ׏!ʊ2 ϻ`0|@Oo\\LW6J[] i}jղoE[YVnK9j* zop9ߔDMtŌS 复 )f7_.^J4Me"?k׫uWϮN"q/qt:.^pu|ȗ0&]*boƖo[*}}ꇒ2r!TXk)|~1Sɖa׏ '~b'}_oެr`PW>x6JAkĆhv;\[z/y/7@ZA/W/<NjUq4}-"M/7cc~𬙤U&KJ@\_x̂Z"!YG]̳Jl$bp.OD2VV'3)slhg0|}hfݘww;uKQ?؛vǕsK /AH==3Ŷ]Zӳ*,V7O\y?muF"<چ6vhaO "KAh*ܜ~.+ş#.4"|:=B8igD8E=~c(S+O;:2,e juAm046S/c،Sww穧zb_1$jJxy >v5teq *DA7) =C'^*2I2X5D!ZD^X+PSuB0kēGRSH#ڝg *DV `򬗱c %*#FT,s59'8S/ @3Y!ઽ2VD''2d?=+x sSN9) gF0:@rȓmE/^au;1yW/#}D , wa^w[ *BV `pykmNAIqP mCJRFZPY],(u'HFbA_EvҌcC3FWJ_.9E8Uu!-g5[TD c8T5]V՜ MNWA'o95U$ʦ2(TqS /h$|(biTx郫Z8N 1%g̑&CwFc-Ŕ-l{8e@VEt>XOg (Yp^  :!oȃruPR>\P!5ٷ W~a/T5{ ʑ!Iu27f]<sLa1¯P~Oq-p@0/\ћΣ Msa5a3Y-X|c'M/缱%Nؼ 4mq;XkR*JzE\Kk\%8rYC'55"4AgډIoƭ&ϫɮq"@n!kMnfEϤ|Fn~?>Ն d9B2;)-զ{}MDkI5Zޟsc\4UFZ*('Verȇw2u#QIM<et]nhgWléC_vſO߿}; [-O[AsGV?(%*FTC7)@z] |X&LiXuۮ8z] f U<\ܨEUV>V)ŌI] p^`,hhDs4#+(B:~U)|9~o~TOS 雫Yxr,F-Q>uKux+^0QzVh,!Li]JF%CޱRXN-VI>_Mjt-wTiL$6x~7,^V*U=[>]1ʷ QrU(&$l1p0kXEaj\eS#5&gϋ3 -EVn7 ymOfV?K/xYZA>1>&rfD)D7U<7qntN@3!PT/cdr:wtZ@޶]>n_Z2e#F- P#G >b9X+og-1^.` lVuw{փ? lV\{̍FW]jZcE JMU5Ba9G0,'DgS<0o4w~7ֲ9.ӪT@)J'x~ٰ)7< ͠#4CpXhg%@qk-zAs} ,KĤQ3J=b&ѪM|o]Ӧ$4pT nhC|.{U2vF]!θS2ràPư1I:|p!Jꍋ ȕ!j JQ!%F#4CpXrvƬ2hB,^D!03FhYU=-u*pWh帱ң1" I>*6^E >Z' ~mS2R@ >i N7S5fEk x_EUAڶ[oڶ ~c\.XCioO8Gxs6rh4b$P2F pX[UjL`Aw}>SC6E5մ\܂=b3% 5򬙼`z="zY̯~UJwA0wstQ=/F+Wq@E} r, {ϔYeY- s7É J (G{it:xqң~>nZ_x wX?!!hF{mpM\'J*GѺ蕡9(rW{x |,|Ԥbٷ*'R\H\F3FuHN>|.EzDռV!I2HQ-ߖJ3q rŲq{`UWnO%U;dA>1>T Eܚ9a*(g3ʨ2_it}>r[dsP< ǵ*`]ܤtbͺ?|1 9?n4- h[T Ԓ-6_ÔgPi|ԏ~g2,ds OôB9?U[2xs4Ro,!/X$鄶PV[@/:zK[7*.~w:H2E(p /DhqQ/@3ydZM1s9`5?]jլe2s†x¡1G>2DM^MVˊ6Ongl#% 5˃婡^B%"DA9`/cT-d:PIenS\"M^/9#aQy7c 1#}$z:h4hhkXcX$2am %x3qh X[U[MPJikWqBC5mU4-QF1]"SqG $Xr[14:<:jɕEO"&ew8xYK@$!гX/cHfį8/LF-e0=^Inh#d@O;WIAJ2" - tZɁ$wUҩfc6ܼArp鏣&%)x0wxe7Gm>b|2zn'QGeA=^ې8sr*7 MLC`5!#TQFǻ dV=ˈU%),[]f-h8 y w6;[&?+VLB^9j(bV(A733f@s̖u*fr:'>FM> E뷓R7KN-񒐒V鈮*2inD$Op0LCn!n@3Y+0i+h( zY~QF@JFhd\pdzڎgMPIu wv4!Ii]*ʋrQ"&v!Io>3 ٛU~ =zk_4#$Q|rz^/҅}Мzo<- fR%Omf?\qZ\jE)b3F[+lҙ6̞Jw\&RgLR] Sʅ?x &.k 0#*>%er}*AIOoy}WA(Q"Dq=7g J洐UM  >C7$^LBye!QBIPr݁Vx ,=*Twqd.YQ as);8B! Vt!ST [OIk $X#DM!(k02U6]zd&n8utWJ@EF-E"D{XuKxӆ: z2z:ӣL%,:<绅JAHWV4q\TBgD)Uaĝo[lxnIMx|H:ђ^%r^=sAC~!aA3fH4,2Y[Fe[+mV F$?W0t-oa3&);;)N\۱ZbͱJ8$_]-0y V!OX`juRYZdKb<PS1}TCRj"Ld7 Z{7,!9dR[7yiƋSSm 4{sMlRmNt ػdpGk*aݏu u6`h^jpp#aY$tbPUzsowe[wpS]yP}K+x:]at%aq12V?K/"ZM ‚ aqeWjv=%DU*;!NG? 2w/?NR@w6q8y@ #'ksIsG%tgzeBf +G +|xN}?bCpw!yZFVGwFKW1F4fy=ot]| q /{B85!wP\\&~y|~z0BKH{EB|&˩Y. ~zo7`sf22Sx }u]3'mC5tebTt?bh <>U$ri_ &j)UzW%1*2(Px)H%ܕ,h"GO-gEQʇ}u&Mw %n%~ v{ tFswG/;ƗKxU4)K=(~KƟSUwkmFJ8` w{, r`A6ٶbYҨ[[lZe:2"EŪX4:p”QtUūsS`|J {h91i!Jۺ|Ȼ5WdzntΟaF0E PTl ~j^`MsxͭѬVzl+`tNm kȧg?'u_kIN':ol M5 V"SLfsf7k4G'8V}(ȏ_ zgę;8A26|U3%:|"NX"yvGNDmY#+=g]^RSSʟKr8N'4.IWr|;藫xrIO9sF0q66}cz΋>}!rW^4N~>8oB(d)?3D!;l"㔤y: 'Aq; P @Tp2y68xBk̭Qa2NxTpO? h}>!4ҠE b1]D3%a SRzXmE#28lΑ` & rJdȟ>4(( qrWg,7D/#$ c ֶ ̓ Aa*J,bc@r`3^ZnRSqhDG`}^ug}8|_.bl&[[rqL[ I+WB, ۹5L[2mm3s/23t-:FdVrdXf_\ͪS.c>}5iUR*ز]) ֟èvYI='0AmH8@G~*y$}G 0z|=O7I iΣV܈naG%uxß}$uWV`i U+b7.G45<},q_i=8gOʈxczT'+ =vmL[ypuELbd& 4h&M3 uXY!̻|2szBzz5a\BfrL忇7xe]b% #!? Tkv>ނ39M ~8(\)?D%煫 z,f`UÆd\.˫rChnM痫~I~i>YLWfNhmOQ}9_8hSUif72~d)jT tƣgk4vqKf0Eb UU8Wˏm7R?s?ƕ7J|4JF]]pohCj:9?/^T\.2f[AձMgHF)}XON"mUYd~bq^ST+I̴osñ4)hQJnrSa,QZ|lXovo&~f?`e4wdTK|}2tЈ NH0ԘK2^{Ng硲Jg0$vd28z%b8m]`?PkQV`t]H%)U4obqb\/*п-|,y M=1⥲*F[Ol>4"#/rU$ouģ8l !]#}t\`0&b3 -UȒ^Ӊ_18)*K@9Ha Dzʡ#"@qI&C]GF"Pr#ϖ~30 dxSuЈ ̠֣dWY VoHQ7zYV"*+A0,462 pл_{|*TDL#WMB6\kJ,q:)6 8@(%$taKH60z #`8Ix4M+iTRZ[}}dH( X=hwvjǝZJ)hևwH:-M2'6G>#ϟ㋄ozyM C5[tB<צ8d\Dx20GF/Fdp<9C^E%*s+ B5fJ6๒F$ƲFdpd9!w=Cx깚w^3I'[,%DD'bPqEp:^  ^ ZF,O=rtЈ N(ؓySI%t#VϤ_&] ofB8.$ýLuL!*$AuCMAL> !E^{C Z~;v9{grLyd8$[A#28,$ %I8^ etZ1K|rRI(EJz{hC2X@' XkY0v LY:hD'+mpGu/by]| )eٮgwݸ*r^E..j6 <7 ʒ/?EV|ɭvkyزͧq0|{̇*qżɹFmʁaTʂK*W7(yfw0goZ^N/=5sa'6,UVkcʣ1Ӹ! &RiqG5R̨^daG/}(nxոx`B3WoGpM@SVibOy5T,6/%[Ӹnhނ[Чq;q Jh_;m1j='0.aXvzΝ_LSQױ6U1j4Ni:zd 0FVnon&||5]Ourkd9o,9HNU u3tꔀwżI\o^%i,#Qܚ&JhۿnoMNs>z>8ۙ@IU)7M il [e֟wGt12)w7f1l\$?Z(_sXZT)lDgQX߮?j_=ksɍwmUmv%[Η][1Ejɑ,jg&Gj3CYK'Qljt4F5+\%Q Cq0Q#S=h(%탭{3(A6uD*8oe #L:1 ͷsEy 5$?(ӊ_qzB;ΒbQo'0biF~-BkR3اCH<$vQlŞڙF~ 5i`=Jöd ɶNl h4}mQ)=C5((]I?F15m("x8ڿg "gv '`ϠՒs uAˣ#k"(1?+"skxȶ"j9[BpQO¤x;UZm8g~𷆞hfh=\y0KFU6걞Y@.O{m-3?AU9#^E}*Le#X0J7k ٵj lYC/fU4N0HƓ#ӈspC2_p1όGp͜/Omϫ̑Yi˻/o(~ϵOXNwqT4&+ tb֖zݛ{Ju)~U}r7 ?ҏ/MY_f(qiD0 VP۸D$mv3,n2.vvVc0 R ã`0[ʻj{4Ԏg7o7aV! ^G||ǟh_LPOmKay=Osv~/#-r&kx:)4]x.^7jw-9j@ՃIS4`xJ 0&e?fז yiMm`8ygLX<GT:&,H&W_> 4Pwx8qcY}<n8XR v7 Uthj„8ΖOv6@6 6\iAݟÓ[_pnoܘo5&~Q|qU$66+w x,S}_, M*tΎb/xj_ߣ'W80+%U6Մ)w lxZ:~8Ks<1`~ھ{xA ~HF`lG܅A';jv4JGj>{/sWxe1G t]KB+8:+%x%l^%IK Oc8CڬvAJ= x[䌸?xvkU1mO7zPՀ>HFCe=j [QhvH4ea$ލD* CɎ&*o-r3prs ò+MBo l{9ܴ{[OcXnO\g/tY_q˄acz VkZ4uI rC2@9a#s^g_?Ћ۾ |۾zT^k[)\ cu~: whq,Lm<^X\Y/9H xOIj4tӗ s,GvgwFDe}yu3xm$pYu]Qk}˗O[{){8 Gqujn9ʅz%]Jyaczo?fc VIv &AREPR fq0ʂKmفif9#/=&dX\?^#oru9x[h)d%鱌(%8.2(#e-daZΈ~Scg'e]^EvX{ʲ}Hmk@"L$AL)8p`KmQ&ўωtd҂5/-(mTh\-J29ˣhJk7yXˊ[ʊ,گEgWEK^rPl}TϪ'3#7,ȽI^{Z {zڈ1cG]v2c} 6z-GIu6"9㯞6"L*[r_w~Үcȯ'ẁѮ x_U98S='\*zKʃQ.{H81"Vx# $yݿG˩A (G{9=UΞlk^p;:-,\>nٹݾ^^_Y1x@-V]HHp12ރUQ|(m6Zw@eJ20ʔ7ۭ;s8*^)Ѷ:ϏYE[`ƂǬ?<[2 Y%m]*IcaAϙ,)g(Xf'TuO$4l5gg3l[[[ !L񓛛8Bh7zmNkst2< :دV ܠiBdX4H.``'st.?3tuL)`$T3q' Ay@*"`i94e]9 !8qm܎`LLSjG h4IIv]6nE,vɠSXmYJVU!}_g`!ڋcj1y1$_?Md &V=z.3A18k(+M,gak쎉!eV gHasqmL.d"[]Oc t/d:ptLBMiR^ۮXI/p+$KIt汋Z`.PBN?!BNpd]J*R)OC[o9~F= B`h399Z|W^n3bP&TOԖB EA;^~ aj0ɴ@d 8%&p!W% i1b8NYFQkJ0C QDxR"ZgFA}QcZS-(lm>dNptL2vo?m1qxрq!9q?YB3bvNNqxΉXkoNK>uд$@>M&Ղ_=vI9+]%*P3E23|J/<0񻦋9EP>2X[X8i>c`Ĩ H4H8zԀcⴐd43V5>N/7`r-pk(x9ZQbjlAXe͞eyCjPXGjTm*Hg#}~h9eg߭Շ!\]se F' ÒR^H *H-)ࠃψ8Zb[5ۏLh*$P@8Μ6;v(C¨Lʐ"xc5Ogouj>SvjV)Suj UtjUY"k?PK2d?@EOCx~wCY`\6lƈua1ɜ`(R{ah^BKb[͝ ŔIx>@~2`[̲V΂MdTH`qi-/Q т DD"&w_J82ho8b=сfcj3m{ Ħ(iypk+UVAT_㫫v?b.|\4y!6ށ|]UӅ1WMz?XzPS:`SW_񮼞ݺy;IY> hẦ-3+ڮʪw~K[q&]g~L7Mhh3ٌfY~ q<P0Q1xyA fX>8ՌDZ.لigj{6_!-y16ۋ:{53DYREɉS((RbA"5g<̙`PԚlvF|ԉϓE/LS8$KBKn묰8{:zv3͋WϯΟ]^A$.~|~ @h0k`{d^{6Nտ\lZ5uUoY5AՐv|ԋp]EE75r<&2^L>aw^.zfl̯S{=&΋H[fh6 ?J =R҄K_~s䨯F|VP+`47H2 gsA%C3#2=i5RVαS2OMPdb@X;{kczw3T!( e59KC ##(1ĉ>Pq 8VT`!mڨ?T&Ŵvfw&k6:hlOf;ecۋd߳qaұ̊Pb(iAit$\xiOvRIz)ľ\4 #a(l#24>v-2!W9$E ;{D"2pi,d"{[K{Jc=Ԓ{3MUD6ʈVSa1q#bPiBAcu_tLǑ.㰃^d>lHK7}u QcM>0W'}L]h6C)fΏa&c[iQ8hw W w %7__n  v0Z̒Ɣ` +"V#u @Jj!< +:-P+Vs "4ՌEV1kWqH: A77 }ƛC+Ҷ' N|ᖪɭ1':SJUiд|g DE$;emzR=:eN*Nz{!Hg\ɍ5?9mkO;L5{7b$Kd4|ظi- $h+#;ϚA۱u2O}'A)NX''{F{ܠٻB-6z3F12Qp C]th9aVs2L}3y'%;-M&KFםf:)| QN0FJNI})EÚ:6m,p'?'sp>]:JƧ*o5wl+˾s= ocOGw.SilQ`h32h$S@"nExOЈx Q*w{W&-w>N׾Ä>;ik7-ES;qtN.*3?PiMSjo}{xx5$#>'Qr/GE,'x"~p" +6 -k[Uc$wp#_7{,by᝝/,eY`f[Ґ`[WQpaI\ g&m/4R#d12dς}L"etB͢QVCfSPcO'sP*LwʾIY){N1!JNH^\xN܁,wWc3.|h6Ѥn WmK:&*{ňnl[܂}I}8VȆ9+ Xih4kscDIJe)Ul4rrQvtLHAG *)]qC }`w ,"7eҚ XB)Bٗz )BDš^‚&<^X !pP(Ŕ8&qςR4ΊۧK t63PrJ&Uc 6 p:C?: 戭%%zrAqIa''NRl7wU7Rj5u<@OKFqd\Y {77WV髩qN$w"<=OZJ^JxKӐ9!I5QS^RϫrKQڙVֳ}.cN}E3eT 1FQΘ`CHb[qB @FөǦ6Y]&`k U "y}vX$ϙ7mv#Uf}fxq\&̷4l[t<7[~Ϸl[ͽ2*smVoYVtz1K7c0wԞijsޥnE'M*i .%]D&IfI7=m$,9Fbw`˜99sNq s8.'nA0nc`1(D # ňQT "#Ŋ@QZ­톮׆m,vb|:Nb+]f}Y-C<9tJX'^Lc?K,|yv7N򞤶i(UEGVl ї-*ߛbզԵ)-`{QV{ئ=\< S Ti,5 4eL I #(:FQ3pܣ(mS8\ν=s|Wuq%zʻGd39Ȗ} zV(p ~l%٣@fuБ,B@@Zqd9A_:0 6X0#iHE"qYK\֊>ڕ%mQmtߴ%sx*@%FO]v{<(3w̿|b.uWN_UzТa[E*:,{|C> 8tgwDkf3ե3eǚpxJq) Q$kRx"r 2g C[WWѸ]c*PNpswS}XKZK a:a#Dńk"ʌqFh,nw[܋v{i?d!)7k:S̅ED%DY"3"H!8#B:Bjs00(v(:qvz+"?xߊd,΃??S͏;OП[Ҷ»,pkWs2xN/x% %q`-#hm!'~02™#>+@ ` &!$:0[0g'۱T9Mr ͅUsͭʓF_y L:Qƚ8_\0jէ8 ߽ؖ\Ux2~N@MM"ˣ_Y'UUu]5V_MYX.ݢ5vAKiPl{ T DrJizZVizZ!1| E%?kz=Fxξlpqс&]i+d5CȑLϨv\mń\'af47K;=  kzh95Wǫc";: C:30'yCzgI{_=sOܓ<HWv)Q쌰۱u2gp7;buY#O^9T./Gܓ@sOAcRnFplQl !Zp.bRg. vwDLrc c0`R'7F5tTI,URZUaTo$if 1hqH\SI&aI81kZ 1/pLuʀH9D['1,V!RY&Y:)cz)̹޼+З= )rf9idѰDC%"O/`fKP`wh7{0( e0ưS7nU Sw4zI!vcJZ_P,) 6<@ҸOe˵2o>yRj:@TGixladvѭ-N'##I|hHdQ0GSgQ0R(o~kxtp2}P{[d͞t|҃_v"-ŧW]9/֐HH)<a-A`iŢJ8Dir a@|IAzWtemAcڹHwD@4:&ḛ4N+!ʓG5Y4nW}, w {/6zYYWSC-ƞb慳A U1Dd]o,;))]*d`):ʁ!"Ryd9aN@fA-0b7RH+RpB[8k8;8JL2kܧ[I;?-L:[xR@v/"c.=4-}μi;A>̫>_qIc=?EQ>t!y)蕝`HrAySDN#lSwbpi{' HIbnS vtb 6-,I66oaܺ|F~$([y`ҽ-#tS O?ДKYr^Z&?;;\Wx 4GH )҉)r^"dѵQ|uzWR:YӝǓ,NXxs5=x 7F|cb:Tߍ풫oA -5)ƚ\jjj=(A>QLߠ߻ewPZ zy'պhaJHu0rq b'-wӖ_BqP2n `9:ߥsOxsL~{7}#0)Elk4\&ZZhUSDSUC}b4ـW&ޟ6ԫ&u-R(˛;v)0ݨ*ϟұݮ vy7\}Ͽ5,tfEzr*w%D8hXG,XeiTf&WÚۘޘL訲|~u]6YyOԳY`ck I"$u:E"m$90H6*F$U/*tiUIΫXb.;Q׍O2(P caAH0)xMLڂ-X$4K,"Ho?*l_#fX#no FϽR-oiiI<^KO*i䡓 N 9NVA}q>*(bX(***#= -̴It{Ӌ~>M$;O @yc0Rf6f8cG/Ӕ.AݔCGV'4:%XwG9OٕSv]|97Ȃ[6)VGjDFg#V ;޼/p;7N:wP݊:M{ˮ-8 &M: u}kH`>阓'thelW=ESS n\ o$wK,}?!qgB\vBa -^ȟԕ'˕GUlmWAypRz^\~1M_;3)Qoctaz"9d*yK>R͌vaejەu׭ r$3MnAז]:{ 暪}/U5ktDi:2"^yl3v=l& j܊Z-VKqNۻuڋ߳\$xfH)ҊgLC3qh&ġ84eLC3qhL(u"s3V90]<1 jc.!fEjV2eg̔3SvfLٙ);3e <>w4Iw4C;yG3h˰ 6|ngFd0荗Bܮ:~C F=vEƷc7 MuZNZXut]^SWZu:}H:RIdo&Mr޲53Kx)eLy):{*:%[!N` kQNm E^@hz{WAʊ!XTTM(U+!{R6+`K3<儉(%*D(#NIbsp`iʪTqZ@&-jѬ3QJ(YuZZ|,0XB󨊈H$!(W6mɪRzɪ}jNVW}AZyӪy:=tf#(kYS=3؇r >b}PRI OJ< ɖ#߇ܕ}sBZ+WnWɕd죞NZ֓9dz6oHvvdk_FāG碒AyiK*P{36۟ IӶZǼiڵPWQ C"ehPrYQoeeyQQψm">^aGzm< }ClW {+3j4jLP=L:e>px1VԸrwHT{ͬge=).ogTg[8blr߭`B m' @OUǼcwھW:{,pۑ-{8W\Q**#qW3zp +rGJFbcJ#[xg$`pzٛRW7Up)~އ6|yU-#:)tLe/n|Z\~ 0  U`ha)1H-h38L%4#Fv@ UEL٪DiO"Y{JE٨61c٦r 0b0V\' NƆ+BXR)7A 7Kp3"5SDcK)r΂wDðqᄈ5񄈰&ϋqכs^q}e ڼQyvV\O!gI >έ1?W]ccd 8 "iL&{FR(fN3I1(吢T(,OK݉;N7Gdf[3nDemdzLIfSv$=€r^_4`x sZD)Z0la%3Xe:[5uԷ0V)ׅ)DuȒI=zǭ&0JqTQ( Ȣ`;rX+?"D4n`BTnJ`^.Bg1RUϗ Df7ߙnx˽eNg0J]Gms{Spߒx=}ɺFxB'1 *֛bɝanpcOfC =scgmI:ZrVDH̃& ^[:^@GlJa^iGjd~sVk/;^sXe3 j=;ܷg54G61{p+͐0cŖ4 `zaf9&ӣYUe,:uT<E,!]Xn5@ĿȈAeL20$Y 1X`Ю0f;ɤL @&F2H"%:(6i[J&uz{WuIǠg/ZyKuK!S1f4Ў94,PFXp4g=Pĸq-©wXu0n.aϓvf|.lm- T'R(_km}r.[5M`†g٫//bfu,P,W_eI4u~1)= ~? ~mY&0]uϸyhɥ\ b_ CTwS;S;S;S*LP`6V[OՔk>( 2$S)nз͵%|eî"OlvԻb;Wޗɤ [4,_F'ᡶ-պN@ϴ^>+Di*7{}*n8aC1TMt(]&Unрc*5 n' T=u.R) [%Νq7CF*z f-p*^j3X4%>Pl5x<4ͪ{E29ֳ(3Q5t s,Byf>][|G(t5̍:1ocCyn=fm_{ABycŒ %VT_]1XIA zX(@S54`9B)Q*6a7@l1c=t? ~}@4rTt}@ѐpFG Jq]q]7ߞO\$G.pr^'- l#=h^cHk9q_ $v-R2,|u>3nl1J!3>|!2A(L% o˖x!!c]~xjAccf]v\<]^ nǼK9H=oZcH!n|79WJNY/.Za\)'М<VW?Nww1Hɡu8Ht\? ZKxC:ogVҿ(O؀j٩-j.s)n?ey)|[do~jxP~}#,XN[z@Á߻C֎O '8ұ}NMT|@X>t Ro7,T2HqJe˻A` !9n1*2w&3 nST ?/X$~c* s'5{&q| r \n Qo<|[X ( 0 S%9lf4XWҙnxP5J}=c^_埝&/NKXbcw@}_'W#WB0)2AZ,aLv*lz?fn/|Ys4F[U7~9O7M<jRVf;kFu Oﺚvag3yWxfVrVy; "rDr,qE/ތg~P`l w]βmEW3fr 10j.J`]o6ZbNMݙo#袽?M䧷l`˝> >ٹ!\y\C"V8A[C@ /l5G娳%Sbz2]Ng"n`%O/@i#@P״T߆Q99}SՏ˫x=Re '6}:eVA(*>Ybމw'Y5Wen}7ɧvkn_iZHC7xq"͟\qVzj-ϑ {g MK{ t>+@O>Hq%XQyܖSh[0犜#4}^(s ^(/PHSKy)ꞇ2߂v}~k$*WhX+d 1s)Rm ցaXqKJ)HQHHPlh w6Ķ"+[~gmU@!ap"׊1 u8x9Tjw/w|a!e3s,N̕skSH[ JS6Bc-xyy>x秋4W6 \ +T o4oial|͗g <<@)QeŪpꤦ0⍢+)c}Xk~^%o.^hv .wa]n# 3r} _Vc UcSfet(<.Dop A[y̑#zLQ VE/PkxJ㧱Ro7#nd&lL=wLُJ; 8}DW>%W)k}G k9cd<4v}xMϬi|ah& oZ,\֑ iL& 'tiZדvJ}b9:){Mˏ]I}K}/#ϑ`Z,MgI7X@uI.0ahp %J7yg) ac{ɒ>:7{[/uAmmyAn"1thQcԊKtlֿenD00,DtuĄ⑉1hbWSJ*;QԝDjj?mk>E7U-qg\m58"Kߏ1)ޡ1w);_1-)ڷ6ít;W ;yS774>q{:o ;x=w84h2/nk YUj 8>}WCUt9$49PLpۓ@gQjtV厚9,=?m.ڝK~S)sI2)/z;{xT )ȴ]Z;krV[U|E/lX{Y{e'Ba'rʟ&O.7 BPX{ i) 󘀲 M(FBȂ2ZwBۦa9Q`<( ssg$Swh5ǝb\NF]ȹ~ޠlU;~] =6uy}T%W:Ϸ* ΅@xh[e.8 93Z{)!Q@`sKinfv^$d;'ƣYx@SRb7P#$*I -"a3-YkaIʾ \Hc 15)fZ% ≲d-!<0 ! :.Dc^i9X!1u pU /^Y ,a%e}X2QZDiCRQSD'rvcF#7!֤D9[r/?&0+Y i&51} &vwqƂ9(@.Ց[kijv:' %1\FBS$HD?o)cJ  Iݪ*@R(8F -+2{ݕkĝ_}a/XAZ`y-`(fNKA^#V(GJ .4()H'% g7MXB pdq04,"sԒ_Buu?f{G-DNrYQz *<4F qF[y}M3y0|71mEUaՈ:4Y7_e7,M5<S5>]~5;t]̶Ue@Eqdt}6gV3@|Y_ସ] X V`D0E|D1'f6dV8c-XG *X]`cXLZ\Z$:]L9[~x1ȗU>xǟ}tMvij&R_j.tvvZ]\iB'V0ͥ"d yOߪQqmwʚ歳yt1ŸoWmls2H} yTy[oareyؗ_3:FRw.!ÈarR'`br|vŨڵ8=*AG],kԮJ)Z:80~2뫙&!ll҄ ~/(6;4ufG瀎vW]~o{zo@ PWhtMAOM@L`c追С)BaTƼU~q_oW-ǭّ]ܚ(}x};٘ltjqB,F~bVjE< Wdץ6Em)hY=g%Y WE@lG:@ wu]>b$og_PԶ/vGbxP#iy՗j$)AJGIT\#\9)l wTX\HzloÂ5=.'A5de`h12X)'iqpE2"uz t v5N8nv~#Zۨ.ҙ}y<Od"{n\[tLhڢW_ҳŦ,B MT 6J!^3Ž %[RSa^W=;NRٖ0?{ǍeJA_62˧ckfuatJK~d~HVwjđշ%Y˰-'dN,;]\ladcI-hpi|PL(!n~I;0|lM9|:SFEMDFE+=R71Oj8\ HoS>oi^[Z Նnq>]n/orGfWdr|`GĎ4mloR[n_[%wD15/9N{u(U&cyP1\22g4@4nN kJ-gaL>=&E {3/.iث5ո36< jH9)P¢'G14;`wX瘓1K+)hF4}Ѱ%wר{ʹ>7T%[}Oܫ3b,Ͱma\R9-p*Yd8@#Z ~,N>99z_?Eof+vy_1V{mWo^On9RtoTڍux]Ɏ1ԯ5vo q{}Mwݼ<{}6{ayF 9wwasrυö*2ջVTw^ 833:R"lHiJ\p|gЮa^IfGɑeF kߪlms^f9S1׋nוy[b^!r >98B.,5Yvv0Yfgz|l{z<?~X _:`4dYMgř}ޕ򿟐9>Z&&%JRXDШ!( Of$ Qkd1 Ci`-{0. Sͻ4/#Q׷K )]ot H'0E]Q(g?pXy混n)SK=; h̞H4!+r6?|yn6zG ~VDOdFJb$a%E`(ɯoSocݐaV9)\d^a HC~R:ܥ?otMVxbZ+,%8IʁgG՗&vupfnT_11w>,~ڇ^tV*jB';/L.`T?9]7 O^{O\ cLvL8Pop0֣n-mH/Z9u\{Hc8|bUI%CkR yl y[elд[FQtMk|ac28\JnlQ; ۃF_D_Kˑ˵ioO_؅ӓt&OޥyEj۝ {OkJFE b_2zto75k7FR&Ǒ $ WvϪmܩ%,d6n8a7Ľ9L=/0 7BQh:-2ZM{So zOЩ@~6xeq{\o]_ƓK,(Q$"D6Ɣ4ߢE'Be]o2ޮ7:pgc *{tc>BC+ f:\Rq(y²?RT8uheYȩSQwPYmh'hq;Z6+msVa-3)\ Ɇ{ܼL|@ k57joZBKFn^0zƺ~8>-nŖ%Z%Z \Lsk-zћ-E Vsh-Lsk1_0~6:%'cN~򄴅r#fwr4&O-ߑʎ#&>42HW߾=nžkĚB 7w .iW-b8ܔ28rwɜxͣ9m|u|M5>/A⺵_B{ҿK IMn|ni7z.o=}p.rѲp7 ّ{.( 哺)]WwmCfd[-Z<.,j7NJVAUUlu},2LBHxf;hW*ire-V_Kp/|Ȕ91Jb>Pq'YLt/O>zo8KF޻/o]I&i"$;'%U_!x4rH&Έzw&FnOKhR]UJY*G9brdT.)|99xmdK9[4ZjQC2E%/L%6* h , F>KFB e%WN&L -k5caY)M95/]msi&#j`8)$I A{v5Z7sZg4^@flYHӴa|d͞stL,Tic"GO@6JLD [TebD8VdƸ>_#x.hrfsK\H#0|>5x׎!/&)q* ZaM´&<&Zj2[$ i|&iMo!x 8IԡAx #. S#xEi -+.Z<zX 9zS+$4) VX۱&)/8Yi"Hh^l̐|q[_pty>{? Br8L2tr*Xj1҉A0$ClXgiX2ZrX[ -r01Xı[Ns5gȭۃ7 HS&H(uicOKhn==7E>!jN)ӌ >\#xnL:SpaM`΄  -FCr{AuL e7ANw0Y.3+H^Xȃpb<FB bk[|LGE0oİB85ZWܴ^ܳQHɃ&!b*x,P4F: Mkۺ tFMFy*9R>pG"2ӁM}ൖ@L %Fh4ZLe,XJgWHhވK_#x>#ˢl)̐cM0se|֩.M`P(JLeDUW#xMm뀗Ca%(XϜB$b>Om֭;x&@6YP/s{}ϟ!H9?sofZ*R:F<R&Z9<`-rd2'ȺL ,(BwվeE/~; Η0vtC'n&Rxg'8Ky$ijw=)h!TvC m墽)Hi/FJ{1^f,YN( eG֩h!iA]\`I TdUlDZx ]2ۖ}Vsy͠iC-H 9jΐiy+pb)f] oG+Dpn lA`F?%PᐲCUσ!)q= H3= quꉧJH'1E4(~D;6QDcr=h>/K1 xt&-w\#2>B1E(qy K0T8p)W l( jT0[K!$Ԧya' :pjKR LBPq!Qaи&"Y,A{dh0 |-!K15 Xͪgtz7 O``KA2y^ r`0t5X8 $JJo8(L@? j8 jG CEK07tqDDD,8(l(1^ET[Ha*F@8#F8Q)#CL>MI]jxOer, fܔ ~%&,WfC?t| Sו2UOQ鏣p /&Bbk0Oee˴6;+ɡC3~Y)Ҥb`^oUyMr[42؆y|6nb|蠢M?oޛɬe~ fTE7*{,/.p1دrbo&^:}S&osYYX~?f.%\~PLcwݲ3 ]\Ņ P8y!x0Sg՘. ye`a,ͭzɼa>䫙_)rDKd^rE2W}wO[hQbf/^kѴl-S.,^;9=}rw3V\irRan&wo}KgBSl\4r{pl{UXvQ}Nfycu]kûL>j$<ɝS+ce7f7u*` ۻOsR;.ןZgU9bO6]5;Yv-3hzxȻk}@K-wC]_sFj?g;qmTv|TS@U7aZ{:Rǚ򛦻;_5ldmvsO Rq^6 wOZ4Rx?wCOtz˼ !u8x>TjEx;ltߞK{߿}64g>oRy܌#c$_wf4JIF?{58ki{8;v.*%^..8?D!_hh:t<~Q:nd_ >s// LQ%`eu&Ի0+ͧ2&M3&}He }ѳT-ÛzN%L26UV0EFTtlf% ,ӽU*?ؔ2m3CC$+y v&JqTQ( Ȣ`;X+\$JH!vfT&$7qX^kBT+&ztӰ:߭Իr|O0BU:7OGS@rgymr|ջ-?˯?ٿKsshs[sY&pmuhW'#׉mgȵԭmW={"%@oBƕe..r;|TK+)o23Ń̷>ЉIMBmm+{FZ ]0#1H%Hb&BEҰԇA CP&EaIur~ P0vY4aA29HysQ b@쉧aC7m"%Uav L`v?o8cVmȸ.V:Z|+2PDhrӠ#֧CGV9RGF_7l+#͟z"KʼnG&wGVrn@%G}w,<仾pk[%lbn&w}dy:`J rԺ,DDꥦ>&` XyQ5tgߦ-HR7m=˅?mΞQg̦WUTr½ŝ|/ ϜIo24Lm-h?WSR*zo[֓`g].KĶ5˺:b⑅bebWS KP;IԽTjװyhOy$;Ҷ]qIF]:al=]ntKt9eӾ>EL;N,}6 32-Mnx)թ|-:(A A=b:xR#sw`{NPiIYItNzPUSѡqxy[;P^* oG|߀\>XvRѝ&<9,0Etvsdfmo[GፕT>h["nc7e:KB<ೕaӳ߫uݎ5K!o^OjOlu4T]?6J}}_^so0wZ`I0Y7Xb/lޠt#%q]?,zIQ MR}ݣHFcw! ר]QQ8{IB9Uaaj낏[؋fvW>x_T}rޝ'\(" HbrK%bdF/cL)D־K=ǶYJ8znXڷ+\Ue;ƟFbU~JE#&h 'H е pXGm,D.^zdL+ jbKFb$S, A9p0IX2%^N)i/ZN {arkU>ЯU+-(da$(argpsM2a3&S D&UOyMN1r_·ybrGZB^\NGHȅ5 M ) ~)ap5l F*%6Q )@8FҎY | \KKȳk;΁s>FuyeZ20k+ޏkۇHx!w|dN L%s[C"Qk#F(`/abQD% Yc ,ǞK27 |mCmC-HwD@4:&emaQiƝVc  B' {rx}d(`ːƾoWXajq<03/ RPBЈY%"H)b{0NbvRRڳqK5"H` G9P'U( HW!9' @c̥<^|d`\ʣNWXJ9aS%]KE[ao`?`" hϒ1y57ƅ=LWMh(//c~O.b_bRt=ɏaPR|0,Ap+/E`͐La@rLASDn%';3Ex}F:5W#\8I ˥CRtm>C⠐O1(ҝ wSn"e}UWBKE#$); 4q&B̗6 @O\SɇJk#Ln='#Y~Tx5^^/"0UE[]r qlZy 7'(3ZFRۆ!Èa-ܟI]Cb1z8ݵ8>*AGm6jZ%G- Chd$ ́}6qm:=fKfxⷪa^_`;~1|ÿ޼ë~D}x??hR8kGOף&P} l ?:4EHt54 M;`h̻\uM%%f˸j>ni\Dϳϯrðp)LdOYеE-_z%ŢPGɉY.k1P\g"SI ,}sHg[E#V>SɶЭ3-S#wN+c=$LgodCKz (LLr`$ \! qV{tHb`I N5X4{GA܆3AWU)FmS (eShT' >l0X}\}캪o{ջ-?-&}gTϭ|]u#K~\g%XJ [%;\P֘Arjv;GCl3*cRb2[d8?(ʾvGA= ~h#-I"0Mq» ,hlH@RO R!Ia'ӂ7a^@{RE t\`z3KbB2 Jht9(&ǠCRzDxk5. z5ޕ-#_ai6Xؗ03e1U5E$_@(7K l*$$*&]tz37YA&| MG Srv&0_⾟>E76@%s4BSAs(Hg8 G*h(q]<ί;=ͩu$Z5!s&bbIb.*N3&H e0"uOߔwgյ1Rxݙ]R;])@ U>Z]dzئuCqS@ed7Q;䬣(:PrT\bBʘNB81x.@x=qh}Ω*0K !Bp u:;'*x}8?;)7ox[0Wjof}:Ч}:Ч}:ЧטxGűr$騨ȫz̓I2jJY^Gx8JUy:)(V ((>^@ @@>bfP>P>P>P#PRpP>X | | | | | | | | | | |ARC)?> Hhu׳|LY9OXX ENr ,ҭ.L^HS48J פ@$$rHP(M~9pُ^Ђv_\nx6,o]'}?Em=Im=_cu5D&qg1Z#՛eLˆ1Ff*96O9Wޕ&Bz W<I[>(}0\òKYiaFV冔'rx/Pը3b8D JQ(qHA~չ5c99׌p42HL2)јC%(MR2}. *pGӢx8Vvs:ӯ?+m;(esC.A!=㇞|CϼH%y?ʄ7ŴO<^6ގh0p[~j-N<뤈##, J2# H:B!Y+1-+AKmd䎚qTxLC#GqB$G#epȐ q98-"vŜK{`t}]65ًۘˇ$żLf77t@)]тׇh\{nwb55ٔl%(G3,'vS<Ⱥl+%<.O3716֘Vnb GE~|h!35<l7X0"k-vuձaQ(v`hcڲ^5NqK.6l AV5bm]7tr˭Nnˇqcqʖ4N ]+:=w\i^7GB>qQN*stOm7ؙW6uۄÌdLDՑ{;h/lS*-92kW{>4v;CYP8x+T8vy޼$)! qk-aqRϤrzm`kS"ET0U䏶t<|9CW[RS6^mޝ|)אjQ}! Gcd\)W_Ek$8!Fzi#dHX`yqS *-љ㼖bDpOQxK(鷟cvo{˂b1EiBdX9!7ْFkIpkRc͘`UEobM%bYnr짜_A+4-77FR#;`oEe{f,ONΡDK1%8ZS+ s(W^z7~>)  JRu$VZAuLGS%4]H y\t JȵE3bqiN0ջy.^˧Δ.g(FGτV()5Sqn ̙C6]-r[Ρ  V⛞~}*owF !z5_/z~n\xR롰oovsѯg[|ㅅK.ž~*Vq#  bnbsC4$K^Й_{VǨo/KmjKPƽYq~d47 ٮ g"Qң"~? ؍dL<gٹv6eSo?Gqﯿ7xOyh!5݀ ծK{Û??}}yݬ?hncԛap'+vXmSq@0>Fe0t$A|1I8̓ #2'4F)u>- R:i䚊׌Kǎ[3'^rO31}54Bg@ˮ!?!?!?ӷZ#B*LA=I5={{{{{;oC?A?A?A?A?A?A?A\c~ebgU<+&2EvK > zPro$Z3-v({p?q)[_DJVc&AƄ 1a=36p*)c'2$9bO1Zks*J* 7f/ O:nF^{>ɝq7nH~|q_|ըdʊyBj8-r ETP|emL^HS48J DqIS4X"6 Q4:/I UDd7}i hY(h{'}?Eɞ$u\M:c0@F)0b9ĭ#JMS wUa>uepRvhek0 &Fqliu,eyrpNAX$g_y4s/1 G:6z5 dnv0oy߇wk=x2*i1\E#׳p:w 1żds{\e7%!h]=걟xȚ{D[QP!D<`./+C%q#ő ImŜ^4]_ q 6f6q:>M ]pi0/9nܱdgxqs~-[,nyy߄܋Eei7{cK)i^n^vvsvs4_)̾ d@gsu|?8@6.m)['>G~ix.Oa ```Zd Ů:6,0ċ꘶,CW S\AK nɇn"Ѳv.wN8tc6_oqMt>J O+BpmQ|ȇ$;3ᰙ;""K{-V,9m nJ'`!I𑘓ki#첎J(  0ՁkEl%x)&/c uP1df(]-g;D%|MڻkRvGuӝ;W&ǘC`Vyl̒ BRBV]ѧhsoRԴ yL.s=ݰK_]6زf9vCGɟfD@+, e_ڠ!ɤ3)34=`^>#e`RJIr e2jEM@H IM.vljc3˜1mJŷ@r рH!P q#E$uVp@X[GZiDQ#lؒ;KbVQ8u:caUS(u, P闿V.Q9Qi&JLDrpus>vZz4'2VCNϣ̹,gewA;ɴ(FIK(DxE8hrZ,k ەe%CHz/15ޞ_ޮ6OQ(B*︍\F4/F$|C)lT" J#,ז؋q`\eg!l52 ,sBQ؆9RIh^*]OK<|mC]T^ZY$Q2#+ r !"yBR== P< UZr'.X۰ <%U| G!j hmJA:b 1q-(IdOӛ1DS eRK$N؄>XT9kO*MPfR{Ɛ`gVseQ)@H9i^rhi2HYIc# JPݨcYy^Do4BK(nϊ*xXWw/??g_Kh~T/5<%~˿6T{* ^Ee_~78HocY3؉&73 \Gq ~5%>n()q;@Ir'~:Mats2 $\5j*+K,&2ȌM. ˍ7D! UY>%tכH}G^-~WB/_^_~TIcF3U`hwMhL]?IN([5j>wf#Z>c}뫳W¸[:q ?h+-mWDdu^-;JWXLVƙJf:oF4OYewژƓz" *)>n9<^mz8<`g]B+A5o/ΈB|=ݏߓ Q*! %Xaw/m鮦SK -˧żBv tϼv9omq)rkIߏKv)[I(\eA\ -`~^[%ڳ8FWoڙ<[@t4u*|oK$!%[/6G_je40߸dG"vZc4ccSJFYʿfQYsBv"SOBuf:1I):6Hm0i8cn$+& ACYd#ǀ/J2XJ|v$O˜|șZĘ0Ieu9t!9n%, J܁)I(З2 ܡVpDRi̪eYqť1u^ۉR[ v4#ʒr~>h9 D=3PW˓ZZuݚ3t0_ xuR^fZۭWw. Ǣ \]YHhޔ)Z9?h9ev!2"{ ޾ҳnW=Brþ}ͱڑa{ `~?OןS- KW] ץpR췔&j&n%CSaa.ËtVX(HڅYb,kgZR g6# pndYZN&AʤA]딻yS+%"Y7Z1Xzk Xz+ tԊwpv&yI0.xz{K՞[Q| Iu51WDuBTUf"e.i݇Aml 7Gv|q'gNfRP$hJ!e2Oβ(eHιӜ:x<%~N1Ku.f ="7.mZbu %1G;=e?%ʸAravyVڮwFWɆWw} oU x2p|VbTy%3מz%k.qk"ͽ۶>,r(p3^쐧s}n]6^^ok)߻ޏh'[v'(ՑީQ*BJQ:"Jsΐ5BhјE ޝ-gQ e"H)qE mz8%z]%KWu.k~!ث>O@Iq\i&Wz ڥMrmqAM>|(b౺|avG!,!|[vo#APJnk "~2, 1,qdskA$j'k,g4 3vqfa){w:gڑ8L~{h|C^C^ARc`RfG+w)AI nB9Q{u&d 7A?MIóv hGEwl47lxL=.#dt\{6Ճ+bco5/)XBQmQ]ð}HvLo;]!rk.~{\Jˈ%cuMyM1!-y_qZtOs/37yIEyZ8. ̳R36(qoZ!X `VUZe]h'XkN ށliqx;rtpy؝Gx6J;Pc^$)#JmG)(F$ӄmWf)t}kzds-v͹oSϹ mK]4oӇlr`(ڃ2\. o߭lfxfY:pMȴ/?rL@6!!*Ҝ,vlwmл"yV7뮺Nw\ښcYiO1K.I Ko5Sj(S97)jڅ< |&aR񹈿]nXƥ[]lYymz3ūѴ*?M ~JkR}L (5]иh2̺,%p:c bX3y}aYi1XŁRHqE )&Z$aQ%>Ĭ"Cc q"10,gL 5)ce4 rR-$472YDRg  5e{H둶O5S*!o` O`Ѷԅ/hqd,x9*'*VWքXIڀHg!3ktw~8qFI6{CՐ%s˙qY]N2-J/QiRޛ>>eiВRYIE 2X6HI(gHڇhѱ-;[N˼%PkqrdVtr}MEi =Qʥv*s9Z_]06a.E֕Y*-BjdX`1-ţ s+NT8xچ(z IdF0V@![ 5BD* {r{A,x4ۜ5!NTvچe)[q]Û ԞZ zYgv0"v=R$c=c=@H)aH.x)0^98ꅡΧf?N[lBƤ!Jv1Aly~e-.i*(8cϕ*gצ"U_ph$*BFLp Fh3Xb#1!Rv:\T* SSFrs%݉\hɡi㿤m[9W:׵94<0smC64]\E'58q{ޭ m/p>Sx09k浢kɔU hk}ҢWH @V酤a5qW6紳{Syr0ύb2@T?h_yh/D^^@cl_$`éFb$T% BPʜ6p" VpsL(QmmudXDcZ)"V_"D4]¢5v6@B$!v mklpeyk#'뷓/=نKOX5| >:'Ibjw G̴ބѸȇϣRB98]'ə*ز 'T:~( Xkc1q+ 'Q NHkc1 ? K`T9F$㑅H>HԔ1тFQ4`A eLDe;[X|ɽ&q5׬U3N.6[VmݖbBtvRZϕq#Ź֓#S:ڠ0^"lHQ>VwӘGAF )Œ ;.Iߍ`) &M2-ټW`ߌǓ0N{\oNQ'l+:r_%L5`X}:8+MM;S*zI4y\'zf%:澾p&|]\3 ·ux-kt:3oEhə):͵]]ཙ|z|E~Jbs ڭ%{gqoA9A ȇ=nNje+ą<[z^cw΄ܔ ;ҷF_}C{ZfQ: F@#4*19]Ɇp-`8s]bޥAOKնKz6 ~3 faw38F츤HQB"V9C 1 TqI[OrRSY8GdqhVrWBY}e -&?]蠝4hV uyj3*I_=Zi~w_ۆ9p `hn)1H_w LPEm iE`˕!@%L`AcCz nqa,YydcW"k/%%\2N,9 dS4 ׶TklF on5څܖ]Qap^:+)OszIC B@\Ha'{FR(fNgc})J56o+whކM6h۝0xp(A |i{FҊrhiQz3myTyD`M\Ừ =]ID8@4i%{6 2q y( @=!blvU`B<UX-zg՘IDk{l5ͭoW[cgZ626!*T7C[q9{k o!,M}9 <1SgN ]OO_O>{>\~ 4l#Ž0+i}N0I:+l=i'MLd$|4WE@,4"#h+4ET8. aQƔ08F3l F*%mF[K\a,mTMك1M[gudb_+S_Ճe`A^|q8.b\|T; JFJ4j"CɈf2ˋ 6ŸO)M&\uT ZOu6&McDފjaGaK]ߖ!,#ٸ6S|z钠xiʆw٠gnp} 'Gw=;ɻ>&8q~tXPbgډw` N?lK"$" fb[ƼQO'| 9t+evGn(=>~9)/ҥ@u .n~҅ԬӺtgrg_a*M9iW>J:ʍ|%B9O1\4J_d2|6ijF-k] MA](IJcgwȗKT\#l);iGG=J*\sm$a&j|=D69}Q 4G2Z8 >Hc+tݥN+:e^|khhmM<04 Onkkxs:FrsdCGN>L+pU翴%g+[_LFHၱ&1sɭ!*$Ť7뺈ǀ#[#q5S8A8M%56 $?{ȍ~`/nn$[юFHgE+vG[,2QSlX_*N9,Iӡ6zl\t&q-w|:pO?g k{tt&t$:9e3(0/xMrzS@RGTRJmɌ|,S^|#iWX0,}v]y FO%uni iiY~4akn$sܞ2mIQe +¼4 R})dŧLQ̔#J_$`éFb$\% BPʜ6ppaHdS ]>o!Uyv78wrds l;|<T ƽd[x}~1TJ(F)uiJ-+pKA= k-Tx55F 2x%02;#Bt!=cXy'a)*Ljd<)V2""&ZH0<Hgن1 tj+'7>r_Mf]Mv1Dy:_M+7s|34AA8 I6h+H% 閬";yt`"̨3h11,%!0Hj@C&g~ ވހ$q}W+_+fNxݯ=3yn]Q+բ}F$N? ~sd %ע4Z`+<@RL#ȯԏ/ΏTG$qZ]/?MRk8G_5A}YoYc!܉=_Fɜjż[jC#^3-咚unE.àrkW{cגz'~er"Kp3"اpKI QDpC@J,s$#9&^@~sUZ0[fYnK$Zb#Jռn|FweJ2/vw/$^0'ku/[̂ 3 $_LT0ma+ItߠdRs?Eydms$M]u^[5au[,~6}!5m[cze>S,a/at;Ҁ' J;-m!P Etm|7mzLQ)MԬ"ǡ9n^Xud Z: ":ס諡†n%@ye(j֟8Nx3md eQrMpDzLidHq嵕8*E g<%L!e1(&2st Lp:oe}̝߫kBz݇ڟ)ԞQ&TKܷCx|UU_|aCN7gI!&Ѣ:Z:Yj&itIUs?!z0Z8Z_@SP ayӃtg7ܯ-zigr TfMELG:o;cƌL{-#chn-x'_F~ZKŽIF 2f<0'TI8pqE$pX2XPD g' qrro)@p3';MvLSZq3ɡ~!h>J(nWâ>UWW.U=Uko}\ȪN%Z|VN9ťoW12# )cLrXb{zB<$}~RUl Vg˝gu l:7 j-$* %^FHHLBDN:`}@oeߐE.qVq`1 @A0 ('ւ=rɃ4&=e7$cZaL9g'1zK밤A9p0IX۲KZ/iR^@((aוU cQ^ g  E*jf0;Zgr<^Vb%VʨP2!6GQ^G|9 1E1ky/~2JFY; ZDl ha%R0#4'(xAc/φ#󔸜,Yk._>݅{sd?gŦz76}OVUGAcɊ?Xߖw?Ⱥ"gtzP% S OAsCc걷FG,"mJIŠ45+R'*.9g4 I˅d$3BH5s DxHQQ/YAooAvifSe_n|7jeVWf#=op$ 9f$һ>ԅsc̱r> c?)2T]"_jY"MὟs#OٚW$#3L5l1lJa^iGjd>9HysQ 9!>yy-H0,WXR >}m(}:*.YϠ}n\_+=]mua43QKIh8q>/îf`߃7@-lN{S\V.)^?n8w])`i}ohл/˘RXQ!y(a+KS >%Gvoj:~o]{<\)L~lSn}quˋI5QsOtٿw|>YϊC{ &4 #봕ؼXj߇&W3'BmK8.7||~;f  u>]d!TMVS-ys:d. 2wy{e_Q׻tיnu1bx}|wx1X7k ,,'H͐Q)N,8R-7i45[Nry(^A;m׶@ !i}LJJ)xFRtKDc0HP8kmuyс7 4Q&B=M8Dȶ"n]68?\--x.rXD DxdVs|5ݮPpnTŒwyk@\c6.kMT fK$H/רB~[L2z%HpZL{%o- '<y;TΌRV Tb< g Q+F0ìj26|[> &W mR]ئIL/ڵTUĊ:y .&nX hV] ~ m *&TW+\%;,̊{pۗCNb<̡~ZxM{iLlg01.Yϲ9Xvaa_AٵBW(si֍vy][dܑy%XvOnju(zQr[ йcYcPJRIcZ,S]Lq+ŚwЖ@Cg ~x_y<|}>_{7((~Ny<|=:Rϋy<|}>[zDVZ\Paԝew<ZZ|,dXRʈ8={RO쟳Rzɕq8XFH"d&JI4(g@(8'*" "Ƭ2(:o;cƌL"^ˈaDk45[!-OW gjeTtªOX)Z@t9 R gUry![ZcL}]麱ݻOUI<<ȝC:4rbu]0Zۻ%Zu9G=.>njVBZ vvyj&ζCϭ!+ni5.<~ M{:&|%Nf]UÚ/|u%dʇm3}_򺹉N+׭Z~ 5[۩Y9;^r0Vso{no~&ߜ*JOWa1^ &$" .aZሿ{ sZD)Z2l 螔(ŸV8d#\8zx%^;Q]I浉JE *XFWQ̌TGCgJzI_\"|{ x.1|qD4"(ERTr tT̍_DƢAQ Aznl $ TR$2A3⩣'sN_"se ]pqgF7b_ r䓈VٛRz~_ϋ b5͹ˮ3I&&vU۫~rOm v+W6}-w*|pJ0 M{d_ d@3[2F/𾽹ue}g]D&XCe+eJ0GKu:?2aU80Q<*M<9ˏSǑVz;YnԳDjiw'8Y~m izibHY)q%xt._$;L_{tSxŽ!q{^u6!8 *p|DΰM؎x=fLx[Ϻx) uD tR׮kJċ|ƝP^o!I;;Ф46zB0QІX9'BcK\UNyRGEjloUۃG uɪt]ݞ8]߶9s;R)NWtƚ%M!?uGrך1ܲ qlŻ}M9?S!pmo `n~s#w|<Ϸ^:(>Xl{O牿Wm([|sEŎ 5#AY͜CGFwV\DYGQ'N}O(J"E9*#h:$NBB,tJ`U R*&dR9clQQxa#/\LFZNp3s_L^;yeaZ n7$&.-[^V[D%(, wZP@uaT< )D%L{(Xģ%(X7DcFSBYNQV,eBry 9 U.Ac*UJ'j Qf*^Yݦu^80@Yfv3_\/y °FZk\h82`$w{Ga'W,诚8 -|{~v?3ϡ?ˉ1M(t8i8,' /z))Ϩ$ZIR<"y.4J 2F ƃjǹ2D&Bi24C*6-=Z23B0kI7(Zq7#U uzXS52xQm.bp%M0V$.x%1BE2LZz8R[ Nl9$8Y܏'CeSk)%x4SUR oT!ED%Pr{.= Eo6dVyĺ(m+eһ$:FCi@ NP9MG餕PiBJKޣ8N{H]U@YKGL APF*!%]\o,E kdpz $ҁ8K8UeĆ28fYJj&[?&*gUj:>Cɟip /?H~?5>4 ]WUf|!a>~oK; M^Ԟ[ A@$|FsAX8O%E2۾({pF5*0 ;%g(8̎ ^ŁifkJbuDN(O+Ko}5/>wyO 1,Jw~nmDoZ.⣆#Qo tiL&͗6ܣ̷0*{iޤy61⯭_7/~>̞NF2e~Ӡq9[nvH L|x|sq c$iH Vt5 FvxX~C qx'Y g=~ݿ[8=*#G-kԮ*f8:7 qF#y`o7#MeqNhfrv(v*L{Y8yK<Bvo?}?~O?2}/?](κ勧M`R x <vhN,547kbh*Jz[x%w2szil} q?̮fi#_r홅f4j>+Z٧|3n|^@%&s{1EV?z&wM[qQs%I- >%RZBdF0vBߓGB qUI:a|~c YҌQC"=%>*OSJVhMLuRȥI*v&va _rstنZKM+pU_:d_})3ڵrFT++"@jպ qQIETg.KveTV9̼xe ճҽ|^b`ɱphpH{CGo/xibC&qo!24s׀51RHMvkS @UddUw]zʃ4.PK˯5d@{?# j f~{Ÿ._{w.K̯(Bl3gO&\?l'ͿWÛ{Oy.\|m$ }Eׂ)km@[ z*l+:ƹ*UB ?T%wk8I"琎T4VJQ(IVӎ+.sك4=Ӛ! dZ%NETDb,I9 ]LRɀv8.E(ƀ,ET&H@xFtoWˏO/5 Mn Z8l#)a F Qq$Wх NhOcdˀ"SP"1"%JfH r2G7'f߮`eY9s+?4>4Djo {oyP<(M('I&q^3NY )A4T-BSF|eiQ;*,6٠OUfp(G(? |[#7h GpF !sxhFEN5E'1J@s8 +6h $= ٣g/~OEHށc옅W!J+:2A*dAihZj$4J^T"EoH!5ip!VΒ̩+zSIO֫J*׀lsJJM$b*khi:%* *7F+BYe⑞7ûeң'= 8ַɿg1yTtQB@=A8ԛ;֓(QBHMR4@kHI!ĥB uP)&B4ed{61$K4M \KX )]U1HL451sD \D,P5rL]-6-SN~ _5o>p*f7t=w]$Nm\ͭ@h}׹+=>i_2NBlݮ?n&w9otfC[mزwn{xsK&E -ן͚fo+Hḳ]9 M;qzR(m_bjªgMSY2MNVcIR ܪn~$ݜqSo2;}CqY@_JXF6Lkހ%At_3 \9@\Yy4쿫csMZ'*P 4'JhAQ$?2:%tZKZf'v 4q/u@j\##]-+9ɖ?YVT|Ϸípk>܊Ek>RV'ípk>ܚBr,1D3x\z]YoI+_f%y C@7wc^0򔸦H5Iٖ7xX$E%E.%V2ʈ"ˈ)S7ZjF)d+GTDzFF~DgHeRíYAKJlBx蕣h CiRYɀ9=0"ALjM<(ÁDVHu@]CXwv phoON<v=iGgu3LGI>Ⳇe|1ڢ V J(.m=W.E+\4Nڠk6Fai{CErYu4R"`u4[m -U1$(T[XugY߀grj XjM >ńWFEk)i-V9nM*b/gtt>@X<NntllM K S+2a7֘Y)2gv49k,TZ0ֹ%IE_)G',͆g:kPԨ=q88)7ᩴ~wfϋ.4)! u!EΉ8ŘR9¼6p3:D<`TERzq|{Dz|[8YRt".6VM!^P;;a몓?)dCuK?^<@DDQ6L8߶W\a(9@Ȥdr$bQ^UO3x|_;E*3@{~>|4ƪ^Y(V]D {N?Xe5([}qߋzFԃ9 8:*&p[O $`֫䩢E]Dc"aWD4eb\pJ'CL}aXwhz; w:6'nOW.U+E,ན :*?qSBѲ .V)¢!qPD*geƠSb(R6*ugf܌R =ubfpZ_ ɘsƷ9H~IGw#?ô~'* >7D<wq5Kpb&&10L>&.ᓭlh&*!aIHeӂC(yT(L!/YdKD (u!@*֝-u~BٟۙŽp5+4uۡMܧ&2^i:E}ײTE;rX'ei +5./BqP$p V40d! 6nGPga2xddJ <(x`!i(U(Ѣ DE' 9=V 2sPo7 .˟;{ 4)CE%y1NhlpBi 8Z'T?}ڟ4"H a {4'ti/R+'֡ 6B0k (vLPFZA*+HPo,E kTpz|b\@` xe* 7L݉?[TyW#эɟ~Som/_^H~Ͽ5:oRAW5pA):\t]ɏ3ʥA@$.|;?n0wVɼ5Bށ3SXrvSh]l4Fx}{>%zG*yzLF:aU>%~Qyl Ҙ;y9yK7ag$S~n?W~n6:A۬op'4-XuS3byBbP'<يD^wǛ[edsA64V 'VG!d$7,pc0Q R'szF;ĿzX邢xX5礲n7WOoSݏ|w(כ~-pѣQ8kꂖO?cЁM]RMc{䨧CrCiWۭٮZ8ˏ;lt7c'?jG̀+~W(׵O{fW3^cWR5 }>H3>Pߴ'S>{ /7Ghqgw$iC%R$UJ C|XH{R\9X@0u{aZ|É߁dI3F 2btsX Z@\Sk([D/ً{OC;Tfsˇu*F/oC]C[cSVd#sxw^o7 ACde@LrsPq1vxCsjf@R"Z&RS`A@FVd5 J p E $L:Qh oUu` `A牖4$O!bz' 13gS0}ÅǛOBGv!lJ8w_NyKoL0"iqBA+}gYܦyتvy0zsxtr7?th@90/8\zas·ǯnc҄ȉrB*0(jI܀,}'sCZ-"S9Ě>(jX~`]:Zo=M{{o$c̿ ]{)ܻPէm!d1j< :;W%Yt9T>0 [?)~NsИ lvG)(CG+<0z% 1jx~h6 h{V.vdl(owX.[ϨCgqk=]ozwgc@X}_wmIxIS5bsPq% jf@36R R)E8B1 NMS[䏦 y e)BNKSE OL35`N6G7EWEp0:z.|BIDL` ܙ!LD$4BhShqqVW3c&-}0WǶL=˻q ϫ秂Ӷv-;3T 2ⓃOSF9 !UIR52"*0pm 5cnN lR^ DSHq4%&%I,POʔ, tz2ɗPyu]éR ^u,<صQfA"s }v(t=\\=Ngy-1лnlb K^}P}O);ԼPrz= ݏy5S%V`$"_pn¨7/gEw ,ޯ/󜤏|i%EWq|sN9|,8Ȩ=8WodŴX*D" u9$_™ = _qd$3W`5,N ]UT%AhN8A? cI #N)2΅H& O5is 2%4Xw;Z<=d4i̛ 8zq~krG]YHЇJ!H . +; $Rq" BƆT6T)BSO ]' ˮ`'W7|aEc"#!-&8+0G89E/$|_Hn\8b4ޕ5KngIK<<ܸ\X%!)OpI.,Q$_7>tWVJQ(Ij!JQ- I 4=Ӛ dZ%NE4Db,I9 ]LRɀV߆Zdk؆Qp9gK0u=,q%;.}ࣅ'\RS>ٍ׸L9Ǚ+Oѷ/7R{ܔrRV1rAEAInd-lg ?~Fj)P(5ZIEF.rC^Ӟ1) AJS ĩv!ă2L,h5T5D[) ?ugGY__;{?a׃ݞd[7CtD*{7Y#5,h*Xh *e -.]:hAleQ 䴳X?iEhu)ۻ)ZcInQ'v ^94Ԕw٫zĀ]oSL(vEkYȲDQ^ܲ(3biD,kۑc$U9,p pYqRxkGv%5Z]lI\h8P?UUV+8'o/MC򏙀>6rXVʡ vςs0|~^㼐rT}s/e|z]9<+DܭMY.yiڢ }? C^fArY?') ~.t~#az\]C,ZXb{@5ce[\#`y:\[ m V+GGP+O԰wݽJl;y+֩gmasm,,ۖqYVLZ܊ tE[-qia{SPጎܧ31>[.]8zG:[iyCӂ4,(arʰLMx5fVhL3س;FUS*-lֹ%IE_)G',.tցPԨ=q8rryx*m MJHcg*Cmes"t0N1T0/ An_HoT{em,E'"bm=!S6:+B?^TQ5 (ʆ jr+ b-LJJ&Gr VUe@=9@Q[ݾnR!~ڼzs`=99W3{7`!QxTSKɵ_r΋05W1%D=(sh*؀JbWI2fJ*ZԤ@AU:*Q(vE$Uy#uq*1Mb.jX[p5yׅɾ9xjr鞟U]y(`vůW,(Q 4wT)|ũOY ŝ=mwQjBH-,n:$NB,,tJ`U R*&TR5cl׌QIta.\$TBNku!2sn.iN>1aqO@e{^w45A+ .5!iuVo/r1&j2IHgrh?.ZZD4Ia"B36)c$|:Ā;m<8v(Cd(]P`Hbّ4_gO9 WsO­Nu7)W|b~ƻ@|,hpW!βĴ Ɗą8FHFCIKO8GBJ{ m2_XF{w̛YQXs0qw? Mᢩ Z |QyA44'Djۛ554%G=eǚvژ[ ;}}?uG>f̻'8=hw3:a6Mm)ӞٺW[핔F{oBB4bFs=Ҍ <7'qa%FhM<,6$m$YJJiau NKqoƵFұ 3W=g~c%5$rˈGSbJ.hp3N4dkuUbkbOn?+y;%G;gmt2LhZZIK@CmoTBz_7j "X+"*ZD)k.lkf/> LRYGyPiOɋiʷUzäN46E#)g6jE% PkM'!&jw.*`;c$:'-*D)iW:GlE!/c>oMGڰ_߻YcˇM*F/oCO]C[71f&ڢVđ9|՗u|YApB2V U&99iqa닷297^3v r^})-T]) ǀ"$Y$s HEDB2(D42v1Z"BD"XXy+`$SaX;$ƭ!}#>.go󓐳ᾐL6:6]z`6.JNzMs,' ꇇu0m-4~;Ywͷ3mf#ˆ뻊/3,F#cF?=z/ok88tUxp ! |bяX;T8Zŭ.y7n-Z?Z:B1 NM淦c*-@ Lu>˫R{R gbe#0b;T Cz 8F J=>XD"FLQS R!OY/Y+;@\{wK>>u"GuofZ\tQ;| N -=˵Vo$y^ VU4"eYeilZZ'#nОȘG/mhe*͈.?MI:q88b{+ ,BYr{Icoo>xLQ9QNH%93ZdM8e)'sCZ-"S0 ,OE28Sz .9־9}$4{QQ"T%HԸ a &JZ0G?>&?B{5cJ6o΂v`򂧠,1z.k~K9|KƛI`Ep^ j4gß5Vm(֝F:'tC8j+v4p_]ܡ(B[s(QZvmt6]ݮnע):b4=AHTWq\} Hdkmc9obw껀$շÂ^L^(6%zxlM3IդN@Ԑ/9Y 9iujP{MG.i<kr2+1LJJ&Gr 2\=HtsU|="%ԛvl_#xf,u{jl=9] V:aߴc׮ |.u3)ݸyW|b~@|<ޠ]u!Bm"1aF"d4O2i đRj |!$hߨff ee `sAqҔ[(U"#DD'wiˏ5A!^Q /CE-.*ɂL\'dh&C#n;j-մ#%%XQDN{]Uh#XKG T(DL#eC-D kpz|$ҁ8K;$, *7f,_W#lb}u?󟕿i8vs[~Kb༇\Ü~ǧޥ|_coʭF N$b0̽@ϽGd}PhdG(#p6!gef f(TϖkGJRDN(OK;ܷR;gG_2u>j1!^{_]obk!R۟R}5鯿U1b!J3Z.MI jie>%~Qymރ5-Zͳ_֮?⍏d$(sW`2skNh<& G$Ϝ4#qyay<:|B q4`ŬǣrgcG}"FmZjY>aK $>w}3qeѧg7Ag利 J;`r'G_~_ɿ?}|BA|j?\GJi Zx |y ;4#D[4ch*Jz{ qjژʭϽn~| f>-#ꓟ|t5"d(+8mdRΧ=ˋxVR 9VrP Axo O=|oeTH҆zKAIP'p>[GNIQ7uFk{26zZ~cDf9h9zJ,U-xHULGC":JT&ɟ۩/9i}cQѢs@\/$~IAYdKجP}zׯDzhuP> xpB+.U>t3ҥW:Np}vH?4=[O HNB$A2*(11fd0*rF.$㋂E,@u"(FE3D g$"8OJH `t)VY;y_Gx론^.m _;~8o'RA+7ACS ë'«H%[Q*EV%HũqIW@Rv0zЎ+iLU6)/X 81'T.h4A4BWx@N(L=ZOY2[Uۓ_2hgF 5޿r`'ujzyvf TW=;i?2^lY + Aۯ4Wn>so׻ݶOzo\絖a:ooi ߃J0=8x5A(subkr"M=oo!Fe5?:#$Wk˭Od+µ>]jaB]*k2!{3.*Os;ߎW;@!N0 xKʁ'],ahD+>1jI29 x̌ *ÔqP(I!F=n#KA ?-llzpZzTS/e0^! Fa4#T'=XIT{f )Ctv2(< S\8td.y6"SdȔK) ϩF \yg-DUf#[8 * ".֫Pd D:# d,櫶 <&$H#"$AÍ& \sDqI(ˀtSc-iߧ겍޵o@ hkQԹO4Л܁k^ 2!?:3vyz뵮.y.eO_t|=oPpߡ%=J = }u3ʠ|J%1.UF0'"g%Ç#qYY֬▸YA̜]qӅ-mp\=d9>_9IF`lBt() ͅfQn}0,R$b(OZ/[UH 뙱ΒN%lSWr4/x'8(KPRl:ac)Ԧb㒶.!eN*r>|sT.3{#{kܷY vDOriq3seM34 /#$U4Fx@csUdujJzgv7{j8歩L?j6{f2t^_כ_KM&63KəH<&FC> S. CD{ RZ%*mz^o|В|0h0vݟ \olVۏo}0k㛎oZ7TĤHH>tCMP@Vb G$˩Po1wͱ)Jl`w{Y [p<|a1En3vg_/kqٻ6#Q`{51~ʄ)J'JpHQ2G%93wWWgd17f[~`hhb MWju@Im6z7'mۇ<>+p=rOI$vn>pfuCR"ixN'Ѫ-5?suNFZ[3~<AU9٠jex&mƼkVtF&mmrJ*Wܙv5udg}mk}2:ڿ]\P0VtpSइOA/Fis&%'YJlr6ORP-*Y\ &xb7F[6ʡ=Dh6=yY˫jש׻TGJ7-8g=L$/u~2yֱro6ϐId;?8mI($'&hFf1@qXgYd1Vg~L?NOoA/S#ӱDs;8hiCogXA{h2ϟ6~s~q6 ^}3MٛƯ*z ^.O.|nԭU?޼zLM6R!yΧ2ؗji7"Syz2i7gZJ$G>A[ۂ|aU320E6$^K=mKP( TGgr/U\anײ-΍1Pk=t&YB0Z+Zx^)>͎qf- lFO 6%!-h,1 m۵PYw>k4:u #_z?dOmpfn[=K8:;8yy֕f07d1~{.dd!b6|ki T#NJhJuČc`PTR")oJZK\l {zGgMrn7zd{"޻Sם}SJU[E=)ƠA֕OZŌ<)+7$?9ed޽JFmMIj0G['F{"l)MN9W Tߒ7p6(7R.WCnTz' DH<:pg77s6~<><}~Xb;xQAh%Y*7i- X:WI"[`2!ER)hJFd =f]29]b.,wþY/$-9OEjFJm޳;`UEic;6Nppb eF]@eC**7Q}lCR;us3 ʊ$wE.!Q*#%(͍6qS}X͎v㴇)֜lܬwu-geg>0YI1a0<_e3zV>ʒC]y?&/BبJܕN)=clgVseQ_Ffc\JBqrQĂu xP $HhZ.:557O&@'t38'0$%{lIrtrozxnޛJtXDfl^ .M(+ys ħڃU~̗SfٞϿkdN1f4S^) 9S{,b1.U&ȼ\S`x 1k}vusek`0 hRƣ3Jwۥ `r2 K1Ÿ3; 0^|y1IFX.wOpҚGzi 6?M T ,ǩ!V1Ӆ `BpS.كZukn{a;XIGR4I8c2cc[;m҇6qk~%G?InL&4vK/ɇ_j68<T˘Zi'f f1Xj }0ցis<9Kⴍ/6YYU}O]~'gXу.j/Wo޼ƭ `_ӆlAs+83F7jsy慡]*ut^zHƫLT[;$bjàIUH.vEVwe٠ <~O4(h~j Fq:l|gqxF3Uwbry]p 5US^gL,\L yc H(w׹hȺ\џ'rߑGgߍGSߍGSv~*~P+vޖLY8kyk,Lig3O \6 moBPJx^7Uqhi{Z|cpV+Gw.=3$aɜɽ@rN[պGn:9ttm"WWn'O?`rg~<ŽYONWG\R\ڕ.ѕ0Y:Fl)/V]5kȦnfTLӅf3@Ze Zu^dpQ"[@y;bk-Op fqZ<2jrbrUgYF6˳^-6z)%sme0W:8Ғ̃w1e Zì !6Te+&IHPH&!,$ o25 mm RҧR)hYȰ\*yX˸3N !=@Xu1s>g|}#QISϞhn ]HI6rQg5ans \a9"s1!J ͊t! ƣ5 :x_ l0: ΎsЎ>K6;qu(~k!cAJd ZR TX\QZbx#pH33 i5/H\ 뽄 & *bڧN`L WrƋ Ў 3 ^L.) a2q DVNvkiG}K&@0xLqY XQ.â1ZR\*UW2DkaNJ=ftЩYE塩"1y6x0t ]WiGM2e>Pt l U$ WJ(2Buvz4IO`_ ,Z2Kfw`A7T" n ԩocVYȁ2aEP]! M`ykc~k@< 9{e0 ì0a\B0ELL&FW<)3E6 D «4,~5Xfp pY AMF@gPC)Cu FT^1CǩJ Ƽ{`DE>1 K༐mpjD&/Hc@8pXZV @F5,x?f؜:h|!?j2Vt- ,%KՈY!$aUCDEr $d9!}oL :EtfH9X 2ը޵q,ٿ".p#@>,݋w&EŘ&,CCFhElsÚszTKKD@KJ7AyA3 vXmAԨƂBG1 $bl}2HeV**26VYg,XN~d} yrx*fE 8-`9%^V0N+*5kk8n Ȁś !u>fds\@7?ZyE r\O.A"0Fu/x `\ة !-JFJQ5z8YX:\cGܠMZ.+_ Xe)iw]4QtQrcya/~?"me]/aE|QX %WM[ OjERHI ;%DRx%T@A t+}yϐ&iizjuߗe9F$z$K$v$MXVR]m넟t;,Kez$o,A[b'Yt"kV#8}Z}w%2f _%^/= vK>?cٔΦ{*PϧgCw#v?-79VG~uxsl'?I7G5ʵ%$\6FgH@jE u}uk>9/sY Vᕍݥ[uz,{67‹Zs۟z6!mo?Xߖ{3.t>8).t- 6/GXd=a_Sp6 !JI \m1hwCjJ|;9|ZqpNt)wly:gtycu3[m%Ș IpN&fQrU+MȖ泥s11Gs g}t>_ցN=c;KϞ.=ܳWP#C̿(P*,c>?9oCri0d<^,ߛoT9% V;^ `k/eJ.|7?V+A-x8+>4EC62 TCʴ RC+-ER*gP!JH]y_`$!}fu2,@QohLuHuv^YYSJh'+>@S2dYp~Nx-*ir4sn'/k#٧tOj~&8o__?~_U dR 2[Yj1> U39A5*9|&sRzFwɻ&_[M~[(4r#]- [ZyvOo$a'ia7CcKӭKSץɴsN5O1vUrV]rN]|kK<1OL~j`k=#IW߯ =ճea6?>l 1'_an;8)0R>nlP;/V(F/ 3A*g{%c|Z.G:TMJ}`ޏt-/ WS.Z%-0ׁpըt8+$[O7R*x^dY9]+תqʚ g=g zW.Vqݥ< V> َw]w OPyb^Ju@L` *Y+^|SfN*fz=LJ8[˲H#:)ehMC]{jNB)Klbi󭃧#/+/p;zpu@c:==ɖ3Cib%\O&Ht U)c'YM_Sa أ{/ŞQk7M6𠥑JVjTIW!1RM,ZnbcKoG3[DRJ5bXGƗ.h~_Gz+ >b>Ԝ{? ~:I%iN''ˋ|+ki͵}OJk0; ͛9oM}@n ru;qb>- .7 GnS g̺E8ڭ4?y%eif03ib|yY WybѤ“b0Ds/_NL.}5Fi6Rdž͓x8=)ێڜ^t8-NvKr/)]YHwӼ] kg<][şe.:;È.ئ'f/<~bcdۯf[y [E Z&MdZu{oJ-nw+ZMܢDKH/qo*$fC)Jw]iPԂ8<idW0/2,{.W% +YEJ] dRR(\V"͞vdeG恩͗#g_7.5Sc7RZǓ_@{%#¦W+k0&^zMʠgRyw1slVt!jô+)&¹PZ:S#kDmWV֔2y5(:ImϏ "tDX\ޤɟGW>[7gU|\}=&WgoK ]ERrT\ƜB+U:V.ke: 3Z?襊AiR{xR68#Q*VPP!9X!1U}5YlMQ tQzT0EW'wr "Twm=6_@?tn[.6-_mزrŲ#% ֡!p3||90\Eګг9E4k5m\3lL=qG7׷mqU,Z|}ē7m<G] =xǶ_>? jՀA h *fQI3N;{I(޽jF&ǘgLFYrA' L֋RĜ5\ˤ֌Z3.Rv { ubiUn3fp uN1YDwq=;?|>5%/+>217aN2LqLk'8e="2HsQ0齐MM5D#l4=6 !e 95r_voܱ=km^v`q>Dub$ &rM I)(O*Mj* 9d2kbbA$0"4ĜYEȩN y9aϒAblEQTX5b"gL\۬ рEBAF& 9F:Á*@>V%m%%l BJIVXDn&a |HD2͕8 i"%?B1]]@Si 1!dR^g{u_ ?o9*+lhkHGl)ԝdP }[M) Ozryѝylu$o9Ǭ-جwk+%V4)HԎl}]bd:>HBC$.mСR%P֐s⣱^wQܫT:~XZ>;rWމ,FTi$bY/W"dQ&ysN!8xV$4M84.I +K0: c ddEù!H!i2H ՗?cPĸi&r`+밿<%UBy.OFˈ)Yu^CrAIڝИ$+ǷPu)Y TڑtCp@bB+Ź2֨ +Hd/8We@Rpz RV`d^ NPG/r hx<{A(ퟕ|t1Ł?kHۢ}m_~j㛋ŏɬbhJщV+rSG 1u''J:;3My?7iyЯQ?'O~h|lEKb0n/Q>MVڻ]8$.G?*J|toi5014Hk=fYޑ4l`YŲ_og =sx1Y69=`GMkԮa%:.42"u(]eq̏r1__v(TO:l8<_H}?|W}͇.ۿoh M )u$YX ?;dL547Z"14W}zk+dSG3; - )~7pJ, TݮI\E] ξBd_LB`RTim͜p[} Čz>h]u&]]=%S޻P?oM<-oyNZDahn *˸*CD{6RD{s,7*RT$iL5'{ 2 -K Rrs 9n ܙsNDzuz `o<5 39if4sz5#ϻfyIG쪚ntK\f/e?Q}݆FӛCB4N%X0JFrQIlܫMZ;I屄gl3e{AsИ- /&lNVst`J#]AF$4 oK9y`@e61: )xkfe 2h熬 9k|}^R/|kc"*'MĘRNPhg @2Âن^MDcAs9h`D2U3B謏@EjN[%XZMYz2nnM|.<^?( %`KpWpH#Ć]4b#{#<]~}].ꃹ<]EJ=L_i-l-po9~?IhS53adU>w=6`]68N` OI#%H;U9sn[0dH:rns豌묚\9w.{ͱGE o!')'mcx X"lgsB o/;u6$lژ\b=+^8525&鈖xsfr+` .)#:#NJԩ&$,6(hB:Tst, ZgQNsv23Ի4k/gipwRF;ylOrۛ*EsGjWvϤcjSߦEn}a|OͯϣBc7u%K<[zp?A"~>b”ZFRKjRƤB4?ZLȵp0.mGJk5FPJ<@<ͼi̾AreQ.m@`[f$.du{!.i0b R&Yb"Ι+2"XƵ֢9Vna' <{ҽlQG4YAfy{/t5v35:MIiŢZG[* Ĭ|pH&E8fs9b b !,* t=4A쑤\Нh$}]"yhugG*Gb M+r>b>G4meN$؅Ga;󩝅e+RtYEnj/Ir _(Je $8%$I&@V\VB Mf5$H|uTupw koX[Z b.x&(~?> 4S$ۅsG03_ /o&M#).zip99 GFH3 ])[y0g\š\)Jg WxyƂ1n66Ǎp>Rg=DjQՍ@ 8h夔y57e{!P8^sff]@5RD29sC Ad $*3!{V*Ņi9Y)_ aL&9@'GYxo00ƕꍜ^C-OSZ:fLyFԒ̝1 M#ՠy|j*i" c9xS{JS{IV<5ڦB<܅=WWD-0wuUU]Fu%@cH]VȃQW\)E]jwuUԢרcAO0!JYO}J{_Qt65:{րѢ&Ƃ97¯m=BB{6H 8H)[Sޓ-{d(+&p.{Kt 4lCvͻ7ps(WBhP{EfkszKDccRq}^4FF)fZXkjO *jŭйٹ~V)Z"̡:2=EMXZJc)u?`G7A:I@.-L%:D58-~#g|w IOՌݼ欐۪gb.'gkjM=w;VXʻݜ::) j@>JH bh땋̳MKe;|>eE3`epYnGf *J,Ef ^g 8KmsV:jL1-w֎+:dBRAW?^ j㟍D6߈ѵ/##H)hU$VDtI)OÑ(#H8.4+{~n#ff9ը}FX |8 J@& Rwz\ )o]vxO[.V_"_|5{ۢ}?Ff~, ̒.䦓E*(4$ii0KmԚ3ęx=b,9A]`%&N@ s^>1al6 OkD l\6\8Fe;l$sdv =dIh\`uBB7, /5L-,*av)DԣLji78o.=m ezvGYǻN4n]_xӞqh23KZ޽pg=֦efYyΖ ڪckv;{bJOd5Ǿٲ杲ҒDN:kWk6ssz 6]yQna[(pR 6 Si{ :LhXpNkWTp3a^^ \%:=٬n:Y5Փ?vO"R( NPM'Z vD)G@8irePE@.2½geɕJx%f6p J ڔ.[q \.jmZXkNkw^և_z 7]tVݜեPz(i͙? YO[?!`)Jjlp6ǚIC{V>sޘ9윷kOͰZ⮦( b㾽ΌW;퉨{VKNCڜFMRyg<&#WA*t p!r4%u@IǢP(4LjeDW^ <29"0i5Z{)ZKCbl۷@R6gJzF4hkO1;ٯm"+m멐6.s?Y<]]_5Zxr~q|JB"2? []CΟԑ0x~v,gc>ߺQݘП;4^CgezCslvcŴs7Htnۥ Q5~J|po1}d8mI# ڇQ,3Gap,f+>o=wޛ{9>*#Gdۨm滋d180\FoQȏC5LUVVĿ|%E yz*.ΐ}2;8@@pFQO&'`x0МUjho>47b=Kl·2+5q0 矪/z㐳n O^p{z8F1?or^o5o}zAU8 1S^/H,@%tcy%1~7lIH %R!jE6 NhcN((Hzh\Սq ΒaZ1 %䣑" d}ƩӁ4XH\ݢNE*V'v̆d˥w~?ei{ tNhZWuK/ >p4/9Oiz@57@bpox|f/y'R,V%hJFg}D%('{JMoF/Ppf7Ks/^Ќڅ\湷/멅x[`;NL_^K"U)uQf}ʇfj#)iE@m5PjN,/l5tWy[[4fvae^^CiJhi .CtR+lQg)po2"h%TIn,$'r ǐ|Fz}SL)UQW\&E]%j?tuT褮LJ}LWevp(3vk^_EA5 "O@/݌zֻn{;!`ҡ].Dy qj Yigُ?o?.O;aXAFͰ]Za{6v(& OKkQJ' iv>h CX~&0݀B2zbS{;\xu֣}쟽$vf%I٩8nUB:?ٻQJiF@ُ~,SwxL W[P@/{݉µvgwɲcŁ~"Gt`R/F0.߀6#"gPH)Ғb|)diGbe&dzy6hZ4[:Uhk?s.xrm`j]KڪktG{8fۀc]D=_/cnW+e=<gcxgrњM,5oH߱0G!bNX5X)è8ʯ_~Y&1˖O|qj.&[~d͸xGyٜMfw56H MM$Qج֛0z_3*tOCf y`f?=CꞬm8(wh}bʕ߃ifbpQ$YJ*Krߘ$O'&t^Hd[i9v#`9]F'Ncޗ $/qJ/'Nq-27lV)S;~-mLe6aJNgW4=EړJ! mH{.jEĊuTx:>$dׄJ`%?z98[\d"weŒeiPzQw0j'4%b@RXoo>oN8GMəKʣ`!zT7QYw> 'oS+R3Eyow1/c7[ 5qk6tWG }4JӱnlB"6r\]rk!WMri1pG|ϐ!*vdkm ;:zb;ޤ8j]lhL3tKH eG -{(WETE4b0LTïݢ/QjTL8Vq$(3\*@ ȭ XEdj"|l(f!1Q'OCKc5_+@ErU0F%N2vH2=r!(X ep`5)9Dlxts*ߵT75iaZSZx˧I+3s=fxV \iSbokK,C[ gЩϪF?~)BNvJ s-tU.@/aN(Esfͭ$G >uJܢS"G{ў;/YHE)D@ R:Dbf4HcZ֨+r XQmDέ]"``f0 D EbLatSY!u3-ʉC> ,~]&z_5fp [k =9 a,^GDa$>Zir& ž$<3SsMra3&WbIAh'qbcƜաŜ6kݯa>xwӈ\X B,T8. aQƔ08F3l F*%xbs)ܮq4:K Xn )wEH +Ҷa5*e"cp(f9#3:Б 9.5í03,* V2E'!A$`{)">j}bZf8 ׶5nMV#\8I ˥CRb،\֭>@He,O+$q)|P"Ŷ=4~_`|t?~S9BR ~> *럂!'?5L!BS?o㣨:Q-A>3Yxr_/^܌\V}bzK;WR=[ۊ(o~.n't3 Gb|Hˆ!ˇˇ2ZB?ʢu;_' =پnVMN.գN&lesDS4u"*a"#i`ޠ(t3 WbR縅^r/(/ʗיku_^y_ËWo>`>\Û_^hdV$e 7ϗMM M&g=^䔗qt[7_^tL];auK٩YiD gr'd_O[+ 1OeԦ]Uy1éYfR/ 5תu_5죻hbT\]l$ݦ'n$)AIOKT\#lѩ\sTgsq%\ۓooD|;Dur]n!Q5 dc`02XFgt8hJHl=u 覿%vx⎣v3T%<97{MhHMz?Wr'wkJ?ȹŹ]Ql Jc1F1sɭ`)$AE=S{ʠgjKg|ǀ#[#q5S8A8Jj  $X'plK185a 9HysQ P6#g=N|@+Ts{՟: s8dNo#ӯ=OБ$ #*hJ"&rw]þ@K"ilTDRU/7mm泽!W;ɐ[~ `݄: i[\h(Izɇqc*g7UTJwnqؐ*t<-(u or+6,rw罶˶ܛwOmw{: v1=,HLnM{t]6KbFnfwuSVߜ/~_{okA{>\Rd>a]@ViCRBC'F>sוM\2mNQ9Aȭ<7 uqEF+ ѲS3Rė{ p*(ISGB2 Ð:ih>AS\`E`'ƾ}ߚEgO![l]"1b2 ]Yބ^7 ^7?R3HδVŖH88 ~ ~( Xkc1`Q^ LvBnOC?]QdKR;ק?7,3s4J+ChBg-CCz(jfag/0U~otIR.߹IA)Fp )KV:֑ !=Vf!Qm_18LT>kR=njtg7>-ҢMw5>LGP<}@.Z#b[;76t9zDW ,Rw|Hb6OdH:6.z4񳾋 Ff-.=wڂ{p  rW|RJ租/bA.ZNlgvfݽ'wIΰU+Y9\|QS*kPL#j^;Z)ƎRq6Go9˯zĔj׏Q?CׁUµr5۟tFqsydVoNAQ`9YJfS]puu^4De5лogwbmDu?3xzV?^١pY])kHUe(H0ǔ(R1FHjB)Šc0*ctbdln5xR}OFb#3UJL֠E)ȓSZSgq(Ej-,l[B UmXC^g#',Qe 0W5i%ZYE鴯Khӟ8t.nR7'Rp$z|9˄o=?֟G~Ov8逸^y/*o>VQ52CtbhN9eB^Z2ZRD^{BR z"euRFel&f˸s-\č4ơ ~?WuWrŞf}[zeY^${z 8MgbfQɀM(BBhEND-|Z]j -M(|W6&TQ ҪMm]lR.$gN1#b͎]ll`68U*(K9 !W&SuC2ŜfYǶV峭JO^DLHEfR]\l8aO.!0El6?vETE,XU)@ҤJ؄ !@t"BA2[cjr(}E ^ a]#]ONjtڙ){饥gQI`"6gE|t^X:͒]nl`x&kw.Nݸ&Vګo~>O޾kp;{3Vݺ-SQs띥>eWTA?-]]w^z۰vwɯKy?/,_w.?p_v2DD'z -]x?t³fr5b:f<=Y@_w=$cxz>c?cb|D.0َR: Ѥj?ޏFlO;ҾB5B*[%ȨHJ5jJ175iu8̯nדvNa.\3wz[rQӅ XA%cdgj> SuAy(erxN M^/ч6_IeDsF@F ˗tC|g송w&W\:c,X]N/FԀ952ԑ.8]^EShw4یҳs3RcS"v$ɽ Dy %L1j B2Jҩ8P) 3da4Eӳ}VP>n=x{mY9;xAOd4O%76XzǟNN~s^xm—"Rγ9vhgqy<3f u;2_(uo}.:VmMFe !&PSbO[$`5Z+,BL LuUɕ!;T+'EWM-X,8|dYeL,:,DiFkggqqzMz"up̿}Ou7n~ ŰGy[;3Zܺ M K:r(:!uH(qL ItÉuWpNN 2 \Aٺ( V,&%x TAPZj[{oa,γ2 |~tF}]2*}Bb\+ڼlmC~b1OW<\ҫyer4(qlwnxqR˞\?ɚw7]\-V c,_'05l5".Umgt.͝n>p|7n>p|7RҺ*px;IX\4pUi8uUJ!iLZg<؟G7_'L?_\}2`46g љ9Ny;"u)& ⿓"A_}^SWI|V"yO)R:b%4^ -!f (i!$7mK`#&fbAA$G.i DI:JZgh&=V6ni+iw3[5u}2O.Fn} Wׂdį[tyn_ FP :H!T\^Ŷ3H- RKr(ɭjtyw ava(0"GJ"MT+Tlr-YK*~YPRdQd!.hɌRƠÙkzÐKƐ8[ԡ]Rz1{򻧈)P%9/aby{Sb͉Eu6ӲUI9z)dsD> 3*&DgH0A>f/¦W9KJ Ե|23b(  u}flB#44%Q /:QV>V I1yDgsb'0%[e4& ]Pf9chO58޺TռH G1RP#.`6*qI^݌cǤcd Y`Q$ɐdfM^FqdUa[phб#ұi5"uC;f(1cڪ;k6lZ? }|3>A3EP)ԅ} L]`TáP,eCEv St*栬' ҁUTXeGA;#+tleMwlQQȭ??\:ɾ1fpYsL7}zg/~B߾ҁNH7.:::b&8-w 9fNk:2FBjŊHѻp"~ͧl\C~!y?WL!#o,v@ZBZڹ=x6_wG+AVk$<*L )Ej m*FYi?'4 N` =!eͺڛ\F?AyZ|>y i7|K"݉,wmmXy,0e~ ,IcNe9Q$dq/abVI\eHU)s%^*!s` d iBY+ѯJ:ixF3蠧z:yiCMd#0?|Jz!I){e$CVeܟ PF# 0tW*GkOF \JNV  2IJ?ϥ\ַ`VԢAgmB6 RԶe׏8^ȖR ^xv[ЉY<,߉>y18U$(Nc } Ƭ [͚ zOx0~gaQ{!!j޽bGaP]=tdf#цnt>:?fYYyІڪu7k`;ݱMdV}A޲V;%:RhurW~9T'mEȗ`uٷ(peh|҅Fi#g J㉭Ɂk^ ?"?_۵T@ Le< dځV$L@ 嘋2[82S>&xqyHkL~*7ܱU)6d|^kH;zg#;R s=Q<d _<>.+ޫ0P~)B2^~մiͫ_0p+tg<_?N`./#ROO.(q߰I+,ТԭJvU)f`DnH)2Y 0HzcKts!prUmYEt\W)"e8?ϯg q?5|of[?Bc%f.+o}S>8ݞKxiyVRuN3"d'S.8(HFwh9FtIA\`ri G嬂E0sB&j[g32g)Opga^eagz\zr=I;zAՠ?xɊZ'ǘmPܲ Cc *a#E,*htV l䃓 @046M`3{,GMNYmhK塠vkcW-6Pck\:tUWRdc[dL;#j m 664' "+:I5,%SY5DF5vٌQ?!+x(Eˈ(:Dq݈A&v*sƸODu*(9YkDl%g\h[R>%X%r19U2s+%%PVimR-#bklFī3W 'CZlm슋e\.vn3$ 4ڽEm"/KJzs'C6x^+gAv\<$Zj d'JF瓊dD(F(4Gg V6oM-=1˅=$q)&HOMuQq¦wӛyms4~cyP)qs^Y2rD]VExqR`RDk$8DTt&+͜6dfYO2'ԍo55mHhf!%NP9L!s6V1!0H0ZssH*.Kڕ:LD)ZZ,jD ]h ʬA@F^BZo@oDCTV5Xw{o40(뗯R ?^jrܸBz1Y2ou =.ͼ0z7[*W ;24mDx/^0]L}1$פ#( St?%G蔦'e: ~#jvn3d)U9Cb #YLDmWS75[?ԋӓ2pxQ|\?Fzz^;׽PH}E̿WϏUy͘5LTGi?921!6?.i8+3UM(wfW%FZ/6=>lv21/?̃ٴoۥEp4 EpF1В%W܌infy| j|,Xfpp=Mug[l}ljY-ZHiXhʭ/'^c~{Ŀ%SCcӛӽ??ۓ?~8w'޿/hR8jGۣk~޶iɘiijo޴tjngEB ]h6fi1 $ϗ_$NjS6nfMpB?!I?Od'4E-Je g!ĉe#ff~Po@e"P߿iMDz`y+e}tM<,h΍$y I` ix4kSd4 inWW!8xN X'1qVh ɃM<c^m6uZ9\}D˹2Ŏq9®wdšiKL[_aRqnR·gm$p&T1)Y)"J[7cHu 㜵9zY;%۫sN='d-aV V h}B ƴ\s敔d٧v]s+Bq8$"zC6dDI4Ĝ *m٦hu/jb aKׯ@HBYXW[·;~=iM`jzt|p@?i2/U6[h-0jmQJ%U]2jb@Q$ZUx#iBݑdvkF{^zky Y. ATG2U>=@BR);b:r (Bt sVn-;j la}a&=ԀWh{t]ձ^kޝЛ^si86,Wmz=jKx0$+*r1XE5:m3v(fc2N CTNJ>6*Z$0YϦ`w8ځ,:qwZ)T{Lo!+V_XZI*ޕ (E>1H\vr=٥Gys$C)WL BCʡat){.5ji%-D֘% z}/$ˣS:>ohy.C/g 8 VsGk,?5ߝtR\%פX3} !2IQ\ Us1\4Q޵tWij ]n`f_sw1/;Fōx,.mR1o>[{'|wwR~~[曬n+g˻ЪwS;v2jCPRKַG]`t4^s :Lw)0j`:_MD/WV VՃ^&-I*< W}V}#.0.nc+xpեt4+Pw-U!Qߟ}9+??xOy87#z;ߣT~!2Sl&E$"'@ (Z\D-O-]:,[l+W'հnb1M}Mkc0lĒmm9K_ԋ^/˓ѻN}J}Q޿.Gpx^C4<3RB*P`cĀG[sHAyT,rpuǞʣdmCxIx'NLv ۪ĜThj =߂`úDEZR"6 bj{7U$R9P0~8{V\w[6onFs6c.z~rh^}fYpHdz}_xWCCk&=jv-vLz3fGn.'ԑǾnw3{5c[#-]=5ՔگDWsj~]tiSzK曥y}C-!XLwa\/p+xn47XޝC@%DQmHmx[vf5 ;\T}} 5@ɻЉahj*yRCn g"t_C"UYzƶ$w-ܞ-9kD֎3N]+=) QL>AT> z<ş:~2.)M11L#OfL(Rp12ϙ~@g1jZ]J@RY9XTA1Phm-(2~Bv5t\`g9זU! !ĮY5Sa=wS..Jȧ/.O>;BCw):OWWcnUSx>%U}Ej QjT\UKKn1Tp_dk c9'_1"Ę&nї)6Kw%KBJ%ŤyN٭Wǽl fhkߦXlW̖Xh1WO..M.;]J@hanrR@.ы({`dm}'1Lɻ3U-hn7w֟POTc|c, ZV@`=h9l6'1N 2;$owҁ;IY4Gw~cIߞ9r(A Uw"H3Xn99j] LƗ\5jJyF'Aa8J ,e k- c1pUQCٖء#v,Q[.9<)' ?3}l$\r $QTNj3^uŒTgVY/<9Xg=36B&HZ[aOH(h١SF$egti_ɧ#FW`T`%SX2֐>N NBk=:"딉YiA<_*]jAg#C'Cz)mٝ^w_A|iw^~ `øV]UD4`V()%,,H`7P_0>ƿ9#}XQ'!^tWBmDj2@5y;hGĻ7}+^[s5@O,Һeu 7hƦ{4ʽLw=c[ɺl'˞SxKoё.5J>9N;i1')PBL8S֔FX8!6AÁ(Dn :l@%`==V G c=FYeb j[Em;3ݚs)ct?ZD%Sr-ZAZ >7V=@@dVU+3Vtqkj3%@.+^y3)W5%FuECAa㱈fDqۊ'gb}(HĘ5V5%Hԫ-a1׌gt5^Jޒk\ԖLYT-8Qgu!:nxƙ=g#TuԢC"l`h5VR45zdnŧCaxYOa;V^=r7(8qFj;n~|GH<>V.Itk]tYڦvwMg?}uOX̘>{3>&w9S`]^ƚX/Qp(ơɦRc.SJZ7ri r7P Ph@$B5-&KdT#ZREhj'N< C;kMY/umwRXC#ȗm ' ng0`l`"K^I_[%-+R˖fETXJc4 њ MפAp\Rw,?۟jiZGAD[HIZ|0U@YKGL A@CY Rb\FG)$J/ ˈep< ̲`Le}]T/٫yu|a}Q8)]Coٻy$!=5>lt\8wf?83%SV.5Xt9Z^?^ Nsqǐ̾qFC ;>B9a'N?q8^fLq>_{c.8peVcd-В洱%t knF67,PhC +#G1Af66{ɪΩVYꬓM6U1ɼq3 dG>W}90YmZ uEqR[G,$=).}ĵ91a\g$=aJ|aODnh9zJ,U- >*OSJVhMLpj]ʦ^k'6b-cz!n} St:4=_B3m6)RhWƗBNR M\YRi%ږN-uYZ;Uۂ:(jpuOU&)Ҕ88I h!lFR'pmNhCK H\kJ=\ 1PԹO$с>ipT!JN3 EQHtY+(rr\,nva4Syj- c ];tLHܶ,kG氚oYlA%Cde,/g='Mb֍} ](n=fK"%j"K:1`a%IV3ɩ R8QĻoxds5EIDc\k\ "8O` yjxd[kO)k7̜6q?w W|( ydjO 糉fKovtxтy&\ 6 ˼-/g0?E#Bb;g:͉WWa NhOcdˀ EbDJ*GfD^:q4ܯ3l;Z;ovI\9|smpcS^ ʼc҄RAIhj|SBhzV鄷ᝢ?MnHX.|q]?o;y@[7~zǃ;K ,qrQQ& Gr+ua%gD{=64ARCE[9yC:-lR5ԃ<fkTanuTZ?~[uG7Uif &j^7`\+#l8cQ/iq"+ݠyÃ^fC}MqSOgv ={:sgL}*` /!#࡜.mX!;W&Y (GE4 2۸eH(}w&ӳgpzɴ\{|)ZrP|lqە ~|*⠘` QXr8ȕo4M}=7 ׃LC':ÿɒMԙc|1+\YsM6gMz%oRYdrPD RqwcGt y< _i^2RH%xbJ'AO s*v ^.jK/g(A +HTȝ "tJT@A!T nV<gm9@\k7. {Z˥öX;Tpgɝ\(3`͙VSEK A)c/FfrZA]f*%+TP+$XbU&wH=udURuu+gYp.r!?SV.Y^EXyygƠ>"/}gq"ش3F%胜@ y/'~bbl Ν^G٣vh!?{v6T)';X%).U+Y`_ztw{hST,~Pi1ko׊B[+M?;b(_a~Rcw˱Ks5gS$(H9w%X#KcC*/gj6M7قee[aMfNkIr+tg]n.B[˭Vr+t ]n.¾[5,BrgtarwX; bg>rxП >*+G}GB6zNB>.Bgv ]n.B[˭V,S!B,^[P+W rB* j\iDe" a0n3P}؎3܈#S^M ho<2Ɛ.h4 ,p-Ib10zRj0(Ebh18o3 RHF%R5rL 9TZ7S@4k2.J~+QƸY ~g~ ޛkoƓx^fbCoFyc>Rt^Ť-C:}6+I$'f \=>@2O5_rK,]?.ҫt^fuNVfU!,d9]5j-xS5/\pE} >h!=K9ݙ [}5gݟ]]!n2Ov66nsիosjl/ZnmL"B⺒G~?i zkq[Klм&M Jփ3.*be-'m2o`cnDK9٫M0A%3 hYb<a,g L#GMUcۓdJ"Wg~hY{>VVpugrj0s3yػd8YB Zp%,x#K5/W:XI;f݅l-e䉳 5$s(@s΂ QRLB$4O8Q!E"BHUW2"JIB8$4"sN_"<5rSw>l9b[cR;ߒ{$?/y&cujBQA*5A#KP2NSQZ *E$:4RYu{,A#M+ Ejx":%KRr>HT2 C@咆'k+be>xK0u k#u}>> @LJέ?%f(|d|N>o~j-E4VѥJ" Jr#g?;3RcN@J,B͠%%6rCEHQ\ CiRs*"ģSJOL}{ڤ(yCJ\yt'ԀK4<[zzL߀Q䔤KP!i䬅ymx1p"ʕF}ckZ3zZ,&C&C $RԴes 8K, &@<Ioq&N Cdp&C(7Y\I1 ,Qw0Y#}Z-ې&V͋glC]&L_U5&deZcd:A.=R }`^,-4L3ˮ֪_F=1Gčk5Vo׬3œuY2@GM_n/=7JS`ijɗ|Dc+'(HQ>J8J|TࣖA㖭koo]!^ ox7M7-͒@P ʠ ,rN)<ֳ A H':g35Чr'DU'qty/kV:u*pB#( }cX Nyaj1>{sMNDPo,ʄ0R%Ȥ)er$dQ^O36v͍]B:c QԀ hN6[_MEl>JXfNl.xA! F%S69k5!baۘ8knkX m͈cl/WV,ӘР98؀H%Q1*k=d Y6:P#eJ EIs,h0<H?h T:@*>8)* uq+!=J5-RŰ Jao +Zɚ.=xGت;UxKFQ5( 3҃N\68(Q 4wT)ĩOyx NՍ"*#ZXtH"Bd 6Xh)`Tؘ8kSn4iƆXX ,tgGO@L.33^VɏL_4iÝ_P~x'8b4.5!ɼ;t3O[K‚7 #HCLYv(.en o=Rn'Y`ŧT~OCb(,U; [HuFڎfߦP`r~qSɉ(ID: } 06Q牳Y3:QOŭg]ԧg#~\uGW]NΦ?/HM?ۯANoD`82o޹88B9][:_ [ތ\̢|A`X`<YGo&N-oխrY˞U1ɼq^1>|8Td8 MM%zp{gA8:??;ߟ eO>3_&㤰L--ogp^iNlilo4װFT4_]ƛ|%~x]=o2f{~>VOϗ_ ϑ׳uj'/8̚h2p=E]b0?ƅޢwҨTW%,Wći.|}OەMnevfg 5oܚׂ+~o?vL땘b ~ zxt}Po˫YuA%s,ձrWmgb~6nܨv `HN?8!$j&9@ R8QĻo6EsJ[EEF&--Xi#1HNDє8k8j\nSq sV}\x;+ WG\K;yw 9 Yxzut|*Ӝiܙ'|Ib臣@ZUi)G{/e®O?y+uk.q9W_?|,iJ;s_=OJnl8qQI& Gr+É Rg,|K9?p)*7gE[ {hNOM[A?Nww3WanrG~ѢοL.Gq @qˇ}Yo-^ ?c:/Ph.h7L{W47 z hѼ2')*l']܂W_wTsxs]̻!k3‹3KKu3KOCpS;WY\!vχ,-•@S* ag*\eiE*K)UWo$(!B,4WYZzB)ہ+J(CpK)vzgJ;m,\A҂;W(`zg*+UִPJhWo 7de!zہLSU. H*djUC‹"Wwh8O ؛}|9kDIYIpHU"$r>ap-zPV|wJbgl,n Rjڭ0|;(|z)V$ # WOwVcp4iyzeo p;ӌVW(&3pťbW*Ky*Kہ+&)!k3p5|W 7 WB8DXegՆiŽY-`%^nvt*rmeP< ~6!VX]/ctJFVpWI(!O`vM͏=Cϵ!M3W>IG_vd-R{5`Xؿ|GMfI"FR7K "F1"U#f\賺~9q ʚBQr"JH5.NL)j|p@WqP;f)>_fB}iUb (t.pXl Ŵ_WpBLKURɈ$Pk/Fmb鷺j/!{ ̌`R՞{۱ˮ.Ē.Oo~beZդEG S#)Ʈ(wygd yu|\Xy켷h*}R=KH^Ri(Cb6:-[>DB`|!}p*S2 J$|MLj_WPe9ջX|6xVIJ8kQXTˬO2t)@ i.iYm:h<ָw[I8jpsyYw [߶,o ?RWzJYd|;Uܷ!JkSwΌ)?Lʢ/>|}4p9wQB*]qIt })ᳵy6Cy6CxjmcGj3t,3r G=JEo:+ VT| Q$4ZQTN=SkKi r9[S (})1Q9dHɻ"Y\J ^ouQ ٟ pBdv׎.^Nüfc\kcb<[lZOƽ&]{!, ./9잩׫gW9[>6Xȝ-j*]w߶{:'\E]ytǣ_Gwѻ9]ve}7ȎGvkm>hOvV|gn}f;~Ct~gLӇ]Njbݹe=ݓwϱZ&kj6_)ٚv /z>V1.;t>A1;*wr t 5U<fQmɎ!Y0 VUQ:%@:MJi( ̔]SgD)$v%!v>()gDa$ecg8m|aώn ߋq'j]%$a(7q۩R"kV9TLL1!`XGjF`!i3S"ZN#pEŸ+X۷ImiLqDsѿ>ү fqه:]Oa=#Goޑ>^w[ѐE8(ipnXBa&,&!ڦ|u1s^zMQUʽ3C?0-M@A.m3J/~?-O3˚TDy=mMݐ ~]hrgy`ytyV8G U7 :}9s?HQUߔ.geIUcY2KpP>7Ces}S15)kj @vir(n^[JW{j޹;xp}R|տI*&㾓Уe}i>i=ⷯpJA8^U^L/3j/ $^={ݸޑtW˯k-]:H["K!iPztd]\_T>t|..J7e=A|Z{(qԐ(5Ȯ+T=D!'g/ e!*9JT&fmb H6AP{ѤcIYUY!j|YIة.xLxt7YOǡ#qD]؞U]PPaS\do ڄVCSD(7'}N@31EB`b}ɠ$R^xُ>f~ԑqq?Xg3-9uc\#.K-kWa;}Eee)ZIɹE%=1pTv܅x۳)jkFo8qFLv7f?RTǍivd_vlHdlСӼM(՟x̔@x|GM T6jLZuҖeBI Vd"/dP$i8'R)t}@G]lZlSV;^cr^Ĝ5o&l[=;)L'彛/{˽^|=TQh,z )|TB-hIhQR2Q@F&`o1PvED٬38Ts(!@A5RDp&"S%Aq ycA}_PxߟݯoϿ9ep~?g~/8Q)DIe'ف[M6 Qm5͚ 4m{x]5¼ʘʭ|??&>{.~V'?U:DW׃P+H6.RΧ=DE_oRseD fv3>P}|%6f}&~F- ^fA+5HCSH|XH;i|Drp@ RIO0ye|8g Dn 4rX ES=M)Y LL\ԺPC5O<ғkS,6\+HE^MY7.ꏿEg &jvHx0kl8绢3U@ M3jV|@n,56'u4~Fɟzm-g:df޼PJpb|)ijk+4ESB Qsʺk)ꖏSM+8O+3חX-۝ʱ3ɘ3ǿ{AGg?p>@!1N/6"S8B m)9(H!RM3Q'c - [e[ heD~v "4@{jhmgnNXި. */⠘sYc^FM~Uo WR߽kimi0l؞δ'g \t7nޘ=y^vym-)u[vHy$hi"K?mD$\$,-H ˬp"3?j0(Z'S0\2XVrORbM%K*aA( &2ΥPHUI$%xsgZiS%Z#}8劄>*卶-{p2<&dS吓ᣁl]#ࣃ1DY~]=/P4'7/? Rg@%Q^Qw𳃟f(Ճ[J{5itSۉ8 Os5`e*C g^bHtr8e&y6ۭ/ Ɂ}KQ Ho-՚1$[\TɓF+x-x8MY{G:q[6Et]YdY2ńjݲDqY玿ޖףE57(P>B 5>' rAjN„S Fx"DJ'897*'xDYƷ?Yf\>QEtI5Ǹ ָ HFbP<-rI\C*zA>.F4cOY/4gd0?Omƚ0v,Cf ëτW5%BQj&XOj&Ĩfj>UŨU9'U&X䚣WHUN\@q&,J.rq,*S+顋L❸zPJs4 D[.2t%+&X W.ՇbuaB8[ګ晎gDQU~Mx_+~b<֏>NN,)-)(Vy#o?Z%1r5:;ne GV _!7IMZ`)UNF#(Q?/O/ɰg}ԗ%H'DBc>>c?BsPaκEyg+&zjNgk>nV7*-wL,RpJQ\JF'VĽ ʕt!pѬ68fβ2E2RJ)t98':̲"@Yޯ\ZM39ȼ"2gD0'-Ky-QXrr %g`əΒ{̛W Aш+$W=N .JT'^bTŏH\-qɥG#2LT.9u+#A/Nê(!␚Ȏ'-fEI"x(S]σrїsYVWN,ݹW1:FRIcm hZnOg-ƞW|n[gu1Zׁp8*|ݬMW=: {tz-ӌw~Sw$#xZYBJÿHg"%4SٔObR :/,(ʑ e+1lq#*l]~^x'{o*=L? j)ziv蟔1h!.hiGV`YJ vH܄rCORbc"K*E,L Ju $p©>J!"49ѥk% M)tH 6vik r莠Y؆'.nB GrTb⳶gg=Eh3B '{<9Z}!S)jx[ C{e ˚w\4ج Φ҄$Jθ=l3.d33l=( J I%hRD"K -#Bॏ Lz]UQfon%LJa$:0:z RΐMP Dqq||kEcwdCބŖ͠N/Z ޗƝU(B\fJ!,][:%5إp6<rCOñg{(eL&YpV3 S7P-K(Jc82\skqP͢*V?39H|d|8ti>X'h9THkCe8Z7{pח:T:z&tYn7"ywdtP킌6U^MF*{ӋjP)^o}Xqҍ9e-8xmTǗaZC S4hf~}srK&P+C%],ìBq}5B>=1*շq+)464^ e܂׳Xl\<|7^[\O˸qS#a[a5篥ͭz@[KvpVz&D`rw D$RhKTNQAT$ut>КJ"4WJ@J,$cI)΅EULbZB ]Hosh۲7)'SorI?Uz֋Zࣃ1 ]NPtK'J)uRNDyFE~v@gaeRíUY%6rCEHQ߄@4xR xPJp"AO"&NhpT(fxP62l{:؍h<OG熈ݮ:{1DyPګELNih PuR5**% W.E-\4N٠ dQZj7P8GA*KMRC"Ee)ھ}jG'^gGP; ;8d_}ٯ4͒JD I3*M,in ă >H1lj;^g77Va4xx@nv zWY3YW}ǒo؄y\,$tQrc0hc: 0Dz4)!J ua/!r$d`x D ڄ # mHȗ2R>YYq%/O%)+U )Rմhl3ꚞꧪ HcZs&qrrc8<՞J1p1\(Q-NZ`ߜhWiH>\l[Mb=<ۻ9»U~x*,#Lr#FjHLNOؼ7"Z'DJ.DiW9w0GSFӋxT=y6̓TD%b) n5 y NAFj1OGSDd JG#qš+0tf0*DZQDFT^D9YơRIyʵ=b q̤'2蘳!ML= ~30,J(|< Gf0"oM¡㥟j4{Zp^7cxFݏ_ngGDz⼨9S#gݨyV/K̏j!Κ#}X QU[@-՚ﭚm峹m|?ۨ{quCԿldps:u}3'=zx*n.?Q/'aۇ޺ӓJcnt3z=}B&I@+R;amς²b<XL>vՈFdGxԈ$^|&14AJ 6Bm@N$KZCsrST4hqE55s͓܃ p9Z_6݄Q4E6_(X[plh2_.=SUP \eDžzA dCwo!L?C麀pY0P|EYbǣ;Nw#S6GmTY,瑫 Z\9u@ma<&I.:Gx%LZs2Zs.'@dN¥G~2ތzxko;PewNrKXbFd)E(*5"dT \ObJ3KB|AB7_@ -U L+`ZKІi&6T82X)8#Deh'o>|=u>{Chfz◟ۈ?@f ́R"A+ %$0T1隌W'5O(kif"%%p9a 8oLI9] ʁ Jɜ#$3LIIVKHljTv"`u/)`PQIrm"񒉠@V=W U+NAC ;krr<d2Exu[f֑@ĩ G*/u Jv)#xkc NP`ˇm /^!`- %玐\嗓(zк `jкbx𪲭(^wC~*(JD5_R0k4x_%E.VVKQ$2̮[#5یғC ĒoD K<7"BwCi GcBM܃ ?omTD Ry;Sys67.w C/`G ./#ޱ^o8ȵ>\л^}n9nMgr=?]ª]oZ@ g&l.P (yQi2@m#z!Her h38զ̗^yT% :$9UhTJb+/E7_L)X_.*H_r@2D'R(TB8K{E)Q)FXBS<nP0|zϛHan ݮۓ.AɵMrp}v'^ C3?p'lPG&-+l9=u}ijt -x4ZCovw3o#[i\- &ŷm^s~.ǟ5xm]mo#+?%wci`p$ ɧbױn#ɞWnYd% ꦊl*^f9tm~]_$~m LQX()F:0_OB~:3:Zԯ9I?7p^xp(x&7{1)yrPjc3ݏ[!۹'~\yy\Ԝ)‘E_Ymvav?ks1O޵+^=B6u\5U6k-wmj7|;,Lxox] --lCsڱg?k|=_̮&ig4Ux~žb~qp3ܲ9e_kɿv~pDpu4pX 2p 8pDph*kF8tL \i.29?>yU2:\e.Apʰw6ܖKT?Sj- Yګn_'3_y7gMӀy/(p⌱3&k)\ŷwnpTFpftfװciV̥W !lW۽'0/ We pXvEܾVh.5= \m W=\#+dk8BvW[N2zzpń4/Hh*̱UC+R+vu1˜:3ixf$qX쨫.\ ף%O(!Wɫ#FUVrWIP!O6rQNxXɤ9%>KGmMzgcsa?[J4OTG}-Op2X p"qm4.$NTt&RBוMD4siV 9cpcgCe)g*3x3/oAޮǻỞ?Op8nޝЫo8N35~Iq(ڥPW-&u4eo 㕧9_IE*V)pZQ$aH=68*q`ow6K'8>eoa EB\=Cy7~^^?܍lQ [uo&vzῬ)a\c4$W|NC.6NvK߶ݧ3H{ZO;Wo ؠyeMpp6U&$Q opEk>e\r[`\qv@uXg1;JSWBvB%3*HB!HpIyॏDʴ J{L.J< 936( qP(I!F-dAKC9Tz^r|17;oxW o@H j:) *gV)"),:OdZY|lH FULk^ K\cA1->%yx5řyEj5 8Oc|ڮ-6=_kg"ث>/h* fDdfvVCw\ ߢDZMO#\ Wҏefw"WqT**VZ@JGERoѴЧ?Jh4IgZ3_TP,R+ODĢF&ƒR 9?bh-b"BWw5$E4kFO>y(\gu.yANC壷_"Dr]ĴYg/}X*au"ʫ}~g~1J@YpkaU͠%6r EH(߄@4xRY xPJp!SB0\0(3YrDL/Ť 3zWmCV61(e7tĆyd!}gn=$7n6x}b' GvYyZ_~XkFB pD:VJv._T~4I߮Fzyѝ"NE>?#j#|5ŜyEjσn$Y3ڋ.\J!v(:K~6U?pf-:o@KܪWt>jV]$]'gM q6+K n].BBt\V$1-oEAmԺE[[X]+s;rA^ ly6~j~NΊ>)l9:fYM]WdٿfVXVW*kM:j[>E4]WNӽr>@z0i(ЊlR(CnՇXǎhՒH>k2mEvRt^jȲu9i FkulǑ]%ًSA״߽Xgw^Pr6L7YM=BN15/-4hI((]DtʈXvS.Н #i|~f:j6h8>1Oe\XQz3ul6ny=| %?Vn8(}l1lޖb*4ӚDE7k:ڶ$=w/f&Jlbnҷ#7gS"{Jxݣ'PQFQ -WAbl/rf:=0PC yHzap $Z B0Q ,rNc+pymLh"OzG2gOi]]~f͍;{d\cl} 9M^&㞡co3TH@<j[#]LVxelZ>=X^KWV,ӘР^98f1aD$NYH& <ZTP `S*Ȥ9-i ay@#I@"ZPdp,h_XPYPc}]MP-|b*7c'Z@lrs9vAI{֟✜'Y!JjJ=OЇ%EQT~ Ì{pkh- XGQSx .EQiH!$爖Pk%d 6BqJaTid,fd,b3cG,\0[ w 4>wls67U7Sߦ_=M^\^ WM&Ƥ xMԁDGK8,&]bR'=12lr^ r ж#$&LF%KW*fvQX21͎]QFmڣv0{(ԡ*mFy eHIePT`ӒJ @+5!Xdګ,X ģQ.PxXs/ Ǯ #"Gu3Pb MPEePQ1&CZC rZTЊEhJ\& 5IKg\p^\ڣ-in Z/ #b1v6#!.NㇴYl슋0.{\\7 X%Ў$h%:є8J,q)pPEa<@؆&Ы^#71;=6nM)Np y? PP$\4@ӻ+dӣ<iXV\8.,8H&Wvz|J@%p'%ŢϮTqjU|zųv)XHڀF+JΈhq"DE'?|Ap'{|ܓ8!Ou?w4FLh6 Jc4 њ UפAp\Rw+|=hib$:EJ8ZB\b|Ͼʴga.O=ɳX4Yw98Uu]>ɼ5eQS rp8xx~lM#QR 5&#P(27kl?n'8T g}906-u791gYoP.|_y_2}W` LE`'0 )o[hNlh,oZ4װFTY(6܏gVlϖjqۻA7rzzZTvg&jĽL2\B8+M9&;g+4j&?h9x]$wqaoDn*9zJ,U- }ƩU\8`׺PC5ژВk]lzLL6dUlt엶c$70_jA(ǰśߣuLf߷ojL=|[<$0ԱEU Q9gi}ǟ&mf{VɋiNgUzäN4wFR'pmEK ][H\kJ=T 1PҹO$с>ipTRevm;lkKպۀKQ[[<'n"䭺k¡S+3 _Skwenw]\;enKʜJг'lb3TB HI-HDJι+)2Gj)x3mʌ_Xq7aVze{M.[Dzrz=,p cGsuߖ477PaE68ÂqPL0]^1b|P_c#W(޵]u98 lOg'?wH?j{޸n4x_QWg{5^lkn] \wێimiiA5kiUW.êo-^ NsG(d%Ho99xiwǸw7 aN9;O5% 0B$TPbrX&JE v$L:{Q~ oPhdqq,<҂F0m9mf\˻dC33{sˠdjú0gy]ȖtnfYb'.;1m= LM^ILWWDC$>i7fg[ c҄RQ2qRHUQ"me61uhjS,3m2}X` Nَ~#(enw(n࡜.m" +NBig'E"VtzΏaтݰzغU9<+kgv KҎRmR1g} tB5K-@ LyI J9p)<1 >9uǩ~(һkmJ۩>,%(at\`R 3B9x`OY혓q?Z~g\C{,[Aލ["f~_.4<[cޖWƴ{V-:l]9-p8eRT %PJ#.aH5q);p+ΆӤz]bYxIF|Lw:|K`=rИjo4fqYZCvAc%4BR| ?pb_ B:\e)U>\l 6 \eq \ei%uR*•=+X\eq(YO+R++7inm޸ X!nrٸ`>*Ut0cPd?8SpR' ؚ}QRZfY59FI5QR|yPfX% YZ=, LFPoHׯcW^ al~\?=Y*9Al*SꅈB>n5Q\f3_)'kX-׃*25.. s:Ӯfӣūݘӟ~ 'u]'ɛ\6:wƞQ2.pxUC7dmL:13n,]s"58Ŷ'ܲF8i5lGyf#4i4v=UDqGpzo Մ \eiuR^#\1 LW(ro*]ei"pWW0IvV]o\bv4_V7O6'r32G؜7؜Ub_9KkĮc3׈FPrg :^q= MQS-9iqԏPNe _<|6(1v;?s].qٷ `0k3~0Qq.^ ?P-P*k\I%(K(QQWMunv)hϴf8}EAHOD"J"1|. &d@tn߸q.ƀ"JʘT&H@xFrrZ@X$\T樂ڠR6( B[(Bx!Q\Iy(, Q:"9,!QFxA)KYMRC!I<-yx˳~y%ͻ_:Ew^i[,}3Eû6-X1vz+# $wmqh}jfAuA g%ōB;Cx-9h T0FMA7?\-JDf{:?=$N`b2MZ8U9ģ=ZM<<.|"^N-Q&tg'P::UԁH$%£N 3%P!³ B0!` P SP&oIԜ'gG<'${>AG$.}b"l{ Jq&prr&e8U sַmњ8!.ojLޟmx[9cx8rr\- Q/|SMB9rQƓ%& ID !O+ȓ@ zq;FXw6(.UzOkOjwjmj'Z5$e7ÚD[ ԍv'Su@wS&KW4hl6IF'q[FJL)8L["bKɓz"ĺՖm LGKVf|(Al:(mQb?9?)r5•oRLy3̦{)Gg(n-5QǕwSF4ןߛ`PYrAi&[:yjx/J ni-&jZHQԡ#A\Mځ `b5o $J2!yeɊZaL'٤ɇuS/ gNt 4(!LR#1YXfDnD0pq:?Xg6)鋋43.#.qq#(2 D,ҁ)|(hc E7(x)F^sG\. fx2!뇇{^ ":> lw1 æݻ&:*@J>~ n2^kBuZ\HfȻph mͣ뉇֊Oɋnn^-4U -h T~]1EMl^By-jn&i g@zĤ7*"^.m6/IX܍o& V'π˓15z ͞=u]e-O6F5)Ѯd̘RIK v0ȈNjh3`3)| =yiO5fLdDZ`$um)Neq~gp1% N6#lgs+¥F&y]ViǬQFYJy>%K0I 4=j[Q~+A@OAvvs{sDr+RXB.~o#2,k (.1˭!(r#H[0Z(Q,1bcO5@C+J74{j{~"0HFv$(fEwZI%̈``"AyxJ@p(VaֆE!)ylzF*GJ-޻`bvRRzTS,8$0μU( ̮ CJsNASDp1Z)Db]ʣNa` S<0',V- #x#ۼB8 FM+keG>+fjp0-B )tg|5b!0o,EINS3gůPɄ Ev~P풡/ Å }},rej)zBa'X#@a5:9D1u'f: Wvxw2^.{T:8I ˥CRbؒ]6pyU*4fUQQ,>.ocj*~zkSHD5lz ήOW/ilO\JLJիMn@O\ɧɏ;5Oi%V!_;onfͅCfzW(uRەH1~I >: af(nm-]5Cڛͬ >>A֕w.\ f^NJV/jۻJ)Z:M˄a>GB >?'.bz?]D?jbQ}p ?}~H?0Qހ,8P 'm,5Ob'1x_Եi47o*֡isu%4+~CjnBVҏO7SmU{5#$+~`U2!BtŕZ.rWE>evZ`~pukQ>iq 'ZR*hNKGI }<}6!fp8$*BFLp ~pFh3}B WT)5,k:U2L̷$+BտOߕO7$r{ὗNi̪cdr. ɞ3?Vz]Y̐CၱXjT ʔV0br!SsA8rԧ'e<pLV8M)L8Tv K lJa}|rVk/;^s%X)*NY.vvI[>o93&nzd5M-"2W1Vc%shr[gp93(=*ֶ<2"^'C:qt9s+9sա=%&$Iftzhݶ ǧlb[ϭIlF:ِmֿͶClYcY]zh󩱲EK-`Mossfϣ]=ytCz>/ҬPogMն"n e_ey6 h/5 [˭JTWSkvx` kk6!Zx3.j5MzYuM0JH&$pΨ`" p%C`13~Qԝ۞$sO4 )36( C}$_# dEvllwFfc{u8./= £J #[SY; Ydm2,d:#;HK*WZ ,q'eR($xG Z\L1j$HccqR9ԕhwV9Ѷww܁ )Y{qj 7` oTx-,qulK#(ې$;6rfǨ$j9 &fZ$4 `xdT*"49ѩ+KIZscQKIR$AjzJň!">14aZiS K:#g:u[ R#uSZ;z;s+'" [#=e ,ϞTZ#Ud  z4L-ӇT ^@/TA/`шLcWZ]\e*)+01iWH0;"qXUVCWJ] /S\IvWKo4`Xouӷa}%p,No>a6I5o=xbNոϔsjd~ o$ p<1ۗXY|A7' /^ԏ~?C0` ` ` ` ` ` ` ``,lsU7l*ۚW՛s|rEx~}r04 Wq)>*S \Q%%,~se/Xe/Xe/X,Xe/Rg[^^^/ *` ` ` ` e>Rk}?]JTmU`Iq,[mw3U~~r{׌n9^>L6Ly+H?-bWZ x@ Te1q .DxІ eF1+XDegJZ{Q1xΒz._!X|ʍh˹^\ًUƄmT 'eK0,s!@YD$C .o.MKng+F +9,K9dX p{Jcd@$ w.#!$ZCL<= pqlߎ1h{ϴLkɓ^' TX iQq9@ :QqIe V'PNɐ)'v4d4Fxq~7ZGwUQ -&ԞhK r|o=~ ~wPl샔ww98TC El~JWNjN$D+"ȉߪ9ԉ@O2~sx~FM2+wՍ9M[Wٵ._&~j|>ͲbvŴY פ~|nY%H ƭ n[AȊӕ# yHaQY,oiCl|7b>MEc.zuӫGedIu\53Ge~HX箯F*SzB_w;]w;U'~tk&_W7<'G߾~߾ ޿7O?G*@=_D1O CsBTWCx9-YO_2W=|֘2犯>nwfp1GhYh•u0{F6 .Zv*&iD[EhEwmݹY/mڶUe&b?[G7Hyo PoI:0ZQABbF2qBߓG cМ?% 1l:;(D vl픐dԜ.{~ƳU)03f)5 jfa @ŝQƜpͭE&EJnC"urJ=h>b"sQ Θ T0QFu:6:#gYw|@0N}kPt;;->ȭ?EbMs}tx8cKcmBٴгg bsir B,.U m9(ԑB3(ϦhFOьs w#͟z ;-K?6-l Cn8)ۋy7xod~:M5v'B7n>/mo͇fYδgx{{5/.|ymk頽p=lNbϧi1)T%_BU^W#7 NjNshoڂ)*RR=RJ<`Ⱦ2XVrORbM%KJq.|P.Jb@IwIK 561ܨW?vrI(!F#4SMQT=*f.Ox\\\ t%k?,dFf.O}Jy=۬ v'LDu;u7>^ &ɵn#goˌ>Hˏe"JȉwAIkbwk!I܆ŜVrH/ai.VZnv/b8nOa[@vb-BvEEl[O G|?ՖX-Q'jōgg?1 R,Pj80F.3$B T&)uy{&Tɩ)wNvs,7t3ɇ^'m۽gơ\k9X+JkUT 3J@J\`1$tr8e&lIoBT6T&b(#1ނ5Hu);STkƐHn':]ճ> i~yWqq)Z^zU\ 2Ftu`Gz<#1 ;VTI)Oؑ/Ԏ4gG",oGYAv:B g GKE%8C T2FM],v*l?=̤xz aśk6 ?~(wg PhcYI :`t\A`:EhSBCN ,!6> I /E0XJejj ڄ # *OJ}#`RhQC ՝ -.Ӊn!xc8ݳ FHN+jHʉ ۥ‡ j9j@S,u>kY%xkc֥Tnj}G-!w-cGm4)5Im J-nJ͒;՜4 nJVKɎ3IW^ݛJL)7/0h8Z޶Dn-ŔT'>=)Bjj4YQkQlm«0#@f:%-#;TǴ ~i_6S$5ab~v2 |rwJiJ_31?f}c Eͮc~}pw7ҙAj^<0InxYي j,3w|۴PŤ5HsٮmaֲwGȱoI\+>R"EJ8IBN k2LwZ ZFL&Z""3FNǻkU1݀c~w\<Iq6d!xD, pY: IyF3/ `D'P#a8ILKJ1,,rF"gB iv9X<ŌS6ϫR3ɮ^>/u8^.W8;' $9D[ʃD{~秒"V~-s ~%$^*ő4{R']La0jZ7`TĂSNq飶ěU,GLJ*fFf͘dӅ$cG]\gϦ w-"i1넌WlsYf'פ_>  F7kl .gzBzlaX"C B&4H%3\xJI[zQ*"$gx6N0+ae) Ӂi]Ib.r6kl;y,Z;t쪵qf{#ص>uXLາ?f; 7A;ib qŊpx`2>\H dxю  D9!< c}9p4 %d#g>YERǢǮdֈ׈F\'3 !@#3,1xkdJY˄J Q&F #$ddQHLd$&=cu7XҠ0I Vc5b6r6kěXmA/NfǴיMJvՋ4^^:G Qd@5RY/R/xGAb8FJP$xh5gJzCcч٤cW}2C>< p)Uȍ'q ~AuE] ls\Hя G?V `o)or;t`0hɼ%P2!6@}a}5ءQ; ex<|87EcYkXJᬏJGFY; $-"rorH ÌМr4йyO Ƴۏh)'0u4\ ǾmףES«\k4ǫM:PmhhWpqAdR~pV ;_yuyL~0?çvs'F?~ba?r}ΪGRw>Vº(ʪq *+@ϿBfg5><{р8n _$ S)Axl6.I$ϧ,C!{;9ϻt?;ڤHbȊ45+'<a LĈSnT97sq 9"rS|`f. T?T+tU!2A0ɵ r B($S1f4j|5#,RFXp4g퉳VJ@\]-F].p%i7nyX^3mmPC.YA=ƿѧOaU+@{0uHAyZx@J2Xߑ>+ixuH.suOeD EiD`%#ƕrUjmEQm{m=Ŋ nfJɴEۍִ|&ѥ4{Ube@JcY$)SWk }nlӳࣦcQw*XufM~ Ʊ}<Vۜ;qTmb*u-F?]t ˜ _VqZv*z'T2LV3$NZI"SVد$*[vV|4A;Eۋz;}S0m\G4&=( xSU0ErrB*f N2~**8Q+DB ~*PJ1xc4*oF8Wgmq E' \Ae xL < )/ 0 Q:EB.؈u"aDza-V8UD.bHT [gh8RRW@dU"WSQWZu5Q{uՕSBW@0ç='KO&{N}De+Y;9+ɨD.JTT9+v + F"Ukul+Ni tJ2r*Qg]nx/S.!(U>urXu1d~wh~|fs̹ .1)pE+h8w}izE6]J%RЈKf*! Ҩ!ݓj?_ӵ1L? L0"NUBYJ%/5ӨE$Nj[^Cqޓ{/us} ens?t>29, #>H};Du]Ҍ+My\l֝woNn&0k@PsnU2̾tgͦ Si> To|^|7Jw|&U0Mr uYa67,ŧW\(8L_fWo ,&_ok`ft3]nIygpOig寋|1oKôxA_~ē2]a*,7Abp9XԕGB nSnHbQ73'lG:ە :;Ǵ-\uzeoƣѴt܍AG6[g/qSfR@[F3 imj!pUx_-Q.׍yF9L[Qf][=G\ /qٻlʌӮKJV^)㥀q~$(~Υ^d_k<)6wuȲ2!KJd hd2Py7 qɛAmr6@KOP<)|1K^&ȩ\@y}X'f=aD>\ػ4IE%Lة`\¾te<ǔޓM_%O%;IBB߽&l6;/; k^<JZ'$@/SaK3;/˫+_ozvklN<螋+߫3j:h[2Ee*ew1gE2(=g3pv<3.wmp3sw%1dsDdZ_ v}v*_QmEGmnEYOûu 6y欨3p&+:ysK9xPQrhwġ.C>Fq]tgGz^2/Y,֔{, jJںנ5} }_Aנk5E n%a=-`0~ˣyL<>.xc1 OZv,o7pZX{R[;Qky@T&d0I({%"!U k@S< j1 pbQPXH}ݽ?;]?DZ@_.<r 5P0NiJ68R$#Y@SJx->ckY3=JNRcX2 ,r[-"P`jȽ1Uk=zj`.:SY&S˵ r B($S1ppqV{1gz4‚S8kOR⚝m1r7{U"d%:6L[uJCGzJV-\P)?[gpRKe k[jUpK0 st2_[C$_pi]9ccђ?dR$3Q:MiDi?BgzqW SDQ5= wbwqO;Xo'$'*'q8ncWbȟH"]H$H(gL)Ml>g @JR!8%E2&k;0;-2h߬m l+ˋ.USӞo`|_o]Wom$K Nl7/2.n&ЬQaAzX^4ÚxJk*~̋]w!txW:ww˻鼜 zwԸE>=A;AZ6O(߼pj;CRA)2n>(_-\{XnN~a$tf2y~oytp* J0\-rχ5ai=;[^sۯlQ~@깯{nGm|s)[;7_o8?7@lny[}T]}\*a -//*q؞-Cl5g/a|gQmɎa4^( RP@=$Ȫ~YPȥl:Q*٩*Z[*( ;QGQCL!X XJBf=FCʫ}Y2'm`|pRM<%^ѲػkF>M/kiXwE/YtF\BE!؅s.d%z&Rj_7Nyѳ.+q~=S!<(˞}ab82 WEdZEI )u  ΀GlgIIH޳c; ypšíqg&˭kÅn[q_VOͤ Q?CGKzel{S"@e6nnIu7?~oizWjToteVz}*X=4#҇{DqMTObcz/12S3OzO#c|ˠ|0#q|m*n@ eBQb,PrFj Ndȼ R4G|hO(}{+]nrpuE]Wld)";X& )lE %Y)t IK!0$"Pey! ?0xd@Yu)E0LJPZ%Th{ 4މMz'^wkWeΖ|

*V1iߪY"Gv๺Y#By7a>()R1+m%RP9+"#pl4lI:bΪxTygo߁,xX5EMdqy2PdQI9<"Kcd dTmXKHn(oV=XCЧal})(vSqf8P>M@ kO411e7v &#uqqusWІ^\]_}Zpy}yc$}*I[j]~?FLdj$&qG{jm 'ugZ~qS|Dcډ0"ec6#6nҏe揻"[$gr6p]9)61*]%{[qhb6T' coCO7'Z $4ÚJ1$SjIQNԒ,Ǿ@HM1A`ϙTc&N]7~R*sKJ^Иq^azIz=_9eY1kc)ų𞂔2'Ml[=鐢z=Z,;|m]Rb1ae\?ѣqmx`m70KCUԑ ԏlB;b|OjU̗b2tqQԙlEٲ6ӭLbد)EvmP(X|hM!g|Q&Am{1[<ۛ,9F ~y:>jX͠#ؔ  Y))`26&k`uɪtLizUc7_l3xcZEt;oN4 q-Ug$+̼Q"yQקeu?]eyVՠ:4ZљD@lCJ`cu:P[ӨUID]Gl?g1sha2`˴FIc*Cp1K`39[*IFNy]`ؙx6|+dR @g8 ~?g ;2!Qmg%QEقfQ`q$ >M5?a (̊.g-O #eNz :6(6VfVfw7l曺׷6}Tw ^b J׷}erd hI`)|%#ЧKc0,T'KA&M*Hk'%c/$/9eT7EFCTB=/T~)D9VG/F_Z#c3vv#c;[6=p{T ,x Is\-0bv/olvuv1_|e L$\$!TzVx%(){1*e2z2_U۶:#U"ru|B6`0Ko'@љsRVgٍ2ͤc_ԖQ[Qhn=B"ujd6EENxgJ@3ڦxX>ҒUPe(訲eIYD(YN"S;wWYǾxDmoY`TCu؄tMo =\do $[K֢1> A*55{وJhky,DRuPr٤ Y^BNĶVk;2edA"vϯ>_2zTkg JbGdJ8˓*< in {GRU Jȁ=&gQXٕ tvه,/?墯'ͯws+_z}tc==w?Wz= WN aе9it9ZͩAJCvy( ד4}c~8úa'e]|8_έrsL Znmy6qC20jKBIނT7IIEO8X[Uw9{dYu7~=GXOKVޖHn騯ߌo梲Ѧ< #إJi2>46Nu*VkF8u<63$ϓi>;ͤ4mķo[p}ӦMKPMyMs5dgABvh.mA\ր${Ggv5*z4 ĆLr#\mڙT(.Ċ_y~|n>L9_2"/HY& uLFscr-!Vr$RH3BsgsRʅGڋ s7j6.8lvQܱ,`>GqynJHVQ"ܛK)^Yi].0a3ȞVXL'nxI bM؛gʸ֟fNs<+`N3;Fʆ68Ri4^e2rSMFrK^.9IC:s KJ&/ж8Bؒ<ȂFDžCԜ4Ѱg_2Kk9sqB6$[ %$SXNNnR|a?Cs=NZvHTxC:5og\\_-ݬh\kGe3׽`U8tmpܡ{BG8hVLť @8eBU]V ugGjo˳8K?mĝ bsMv6ͷW8??r;i[i_LN[q+)9c>.Z40r3lwYGi[}?cxu/@,#[VY죗Xg; bX{ aU?l]h?EloOݦ\M'}/7]t^S|vyyq~ml>MΦ Quy g﷛Od;#s#" B/\ 2$qaDiPTF@/) r M\ְo,lpp9v9S?V*[/͟;; kC6֭_zrdiV-/,O/oss\@"SRBny34|x)>W8m rU۾xaܳ<]em ]v_V[n1^}}{7܂߇Ӂo.ua彛k@{^ ;bxOĐ1fp@VuBꁆú5^ű⎫R #Ɋ=(XFޒ)Y۬YAdǍFH;6 +9(0_t"s$ Ek))&kAcU).T>JsKQ.WB\*HE"Ykc!/6L+ &УKz nQ 28ō&y)9V9U𱃘1\yG>E>?zh2|p4VۈTf9gG1u2Z:~~n)9mK"q)Y͙Rգ0#)18OJ~ b&i# CP.ې/2'IRO"{+3 Y s48:N>~ODӃ3!vBnF9FmYD9M_/h6ɫL_fʐQpf:$S%dO٪OyزJqU χy/BfA J%{nRzUlsY;N&zo־vq]՜K+v% {~ 60Gt&F8IfG tuzV%:GќS5x5>Y~K}tFzѬ?pL(Q24̨jc6҃8¹*ߞ+gMm&Mԇ__̭fBX袘^7#Dj/so6)=m|(ujL5qz*`iݽ\lIcIc\ރ؛h{M<"'Q*.4=+( }VөV8qy|=~;o^LټAƳ9j8z>O 5~y=4jC1߀ObO_ o(O86 zh-MuM?u{я>kMpᣑQbpYpiSmJ[޼Mŷ|ym۳V,ozeڂ$_ R^-<W9R:$a1uX.eD <󐸊"f B]HI՞vJ,CaJ .u~yCnZy. YpiոP~_ɒXne%~29߫Z=ZʂDeힴUKz3i^V1wptbo%6.:;uW9%˦,e*]"Xsqmũ!f>$J zb=NΆ I\wg&(9Lʼn T9iS>FQ>'E2\6 ǫ=7[AwOd%IUX+WZ2yo# ĵ<i3ֱ$wg~g8GĮSPeK-T-1A~VRXܡE\t:ϟ\e&WITp!94oѹ?Nj{9Utv֮gHi}D\i]lв&cW7lK \2P|-*8Y8MC3uƊzݫ[4FpE vGr.o0,=vp3z_ >#59jlKu|e"bR([2hL \GZzAj*z#m=4K78 fڷ$|r뭯Ayԃqעxs( &ÂJdtP3%+'Uʂܰ}q"@fɒ cR2b[k|I$E]̉S1,io2``s%$:8[=7S nHS" KPlg.e1 _X&pE9EDJK[LXCJkr9CCkj`9adH8QHԂ׭4'SA tZ+\ntb BjSJ HJeCNeN>3VMB$ݧ.]'B ƌi,Bhm,*j!RR#N"֕g:Fx"\9_>ݢNN=]@e-u%k UmI$ /xS`t6ČU)S)֔d4Ud~l.U=ep"\3 nD9NOC.3k\D%MLvZ%]JE%~bOnrS~W^*rr(Ebj9bC,#Df,+ZǃGk-DnaJpgC\d\%T5Qi'4.FĒG! Vx'ռ"Mr3X3`4ize.AG3^BNv9(x1)# ⶃ2] *Ҏ)L@laxLqYR (x?0F3`-ة携leGVz HVIZƒ)tC, "1AX1y6l0t }[:7 C[odE1ɐ( R]pl#+0 AQk`)4jDÔJ* kPЀXOT  B/^', R؂q߂VŢd2C]4<ϊH8 4!d@kOwC ;DRޕri `vtՀl!!~(3@Ơ) 7* bȇvs+A aeK8h2jb)Y^Y @F$]@8JVr ޔlMp%BA1WXU+- : "</:^K zJZvB1.%%uD8E>1i`p% IX^68V[5" $Ɨym$Gh"*((C  e:-Un°pfcuk?f؜:@i|?ϴ(wkBQ K+f櫌EErٰX̉s!sL L*RtfH9a2UIo`שX/bּ5-II6 S# hX?P{2ȋx Z@2ցGA8 VgESQ_ ~ S,0P 6b$>I~aWVߟN13 c5 .z+hP6:=-h=c;&*xQEnTAujP̨<ŒJ[k! V`:-(@u"uQ82 gu]~amTgP'0SҳBiy@?xZ#7 .' !0 1@xг@0ND$EfBIQYkutБpi<3`$k3o^ \)gj{H_!S.H" a b~W1ED^]|fF"%qH%8bwuuUw=OeUƫ ؽ!vZzߨn039F@ 5CoI Ë!Kָ:8,y||':]i9m]zj^.y#T`Q릫2LL[PSuDQR԰¬÷5$wy.j [V -b3pƸAY_ φ`na!rE \ Er v*z:!0*@%Kk0z G:~V lqw= 5Zc4X 'Ya_r\8p oaZaZ8 ;"PaQÃRŢ4Q50g :pU .c ]E ѳd|NiuKkLr#SÞ@jF+\#xqDEY,J3SD2AUSEc ~s-c|xdϏ{jyGv+śZPٗ9%* ZC &FxXX.a0Jf-th}/j@lex/>%n#X>C Z6l!& *V:"]v)&#`f%xZĔ%H"MXrM.D]lMQ~5 oW$hw*;c"DoaLX`ઁ#lꅉ?b쫀 V!di)5DDɍiX_6?"meML"40qtv3t js%Bc 6q$A( :Z@ -g#PJn @mՄ"$! DH B@"$! DH B@"$! DH B@"$! DH B@"$! DH B@"$! DH B]$wൾ&$Aھ$)zHSyUéJt3O?.y_^9//>͛?n60/U>C Tj̗Z;2i)Q"&F񚨚 ]Yӳ(tO3*JʚG Qjvc^m)0(h T%uNQ*ˆu75!`0䅤}e3[REU*𪝯G,)k7%5L[#F5­npk[#F5­npk[#F5­npk[#F5­npk[#F5­npk[#F5­]f[ׄ[Z5&n u(֤*XI?E pA@tY]ttl97??_ (+pFVa^OK7>8(~DܔƴFzȭ h]wX94;?K:p@y׳Q<ܡ׳QL|u=8`ƒ!#pڿ,۔N@=Dݳ W7]M{!saI,mY0V<D}s!Nq l|ۯ,Ka!^RU O#لt|?Zv+ !ܙ\w|ڡ]6i1(}oj~.iW4N/z!6hkhwrW;OtY 3܅io_,͇U7==DOo{M{ jcŜbdڅ$8sZ'(rGeux.f\qLqLqLqLqLqLqLqLqLqLqLqLqLqLqLqLqLqLqLqLqLqLqL] =M|"z8&@\z8&@Z_<H=qLE pcboz9(L)QppFMGMbe/QXN>OWykp6ga=Ϡ9b),:Kf,Wy{Gl bibA1Spw8=kD˧F]:O{H;|ܿpb$Sxcg0ϳ.\ XyɻE&x|ҷj/0CwuC:qr{R EOeamN&Es݆P]1CwDlnǛUi;t< aeHMԇ>!}0nS&zkHM2 Z\^mZWx!5{Oe χZ0 KvVNˍL:ZIXMBΨ͓_?߭a/i9 ڏ:j %۴5{զK)c(Ys4؝,Ŗ}q'io;zڲzUkH+a}Pj[GZI{XO5ضa찶rap+5+qv7$!EjW#GRI`(0 9G۶LLw͋uIa͞Cm,'om E}mG =^)4b!ϓ'W.nA3˄_,!?fSM4SǍ͗{ڼ}[YK_|f|ӫԿ_PA{waoe&n6pS۾ö'θ'g(tH!}?r\~8ݽR2%0S"gc= p2WIJ1>Te4J[k*VG3]P%!1B0<_r:QOj5ȕ,"%KdRR(,1EV)QdfYVח_hF;M7Ͳf7qK*ix[ sJX'1IJq ָTRy\*Kq<.ǥTRy\*Kq<.ǥTRy\*Kq<.ǥTRy\*Kq<.ǥTRy\*Kq<.ǥTRy\*KqNyǩڽV|SM7Nn$ New?i*_PX-k w -/k쫩 :WT-Vy6z5ߋ<>iBqg\~G(뫅2Z֘YhMoԔͲZ!3=J a{t< ^ƺ.N'd( /Rib]o9f sF#~DhXr~I%80*fǷBQri+U,!7+4lHLU_M"$[|ԅ:Fш u4q/p6W+$ W7MZ\޼zXz B5櫫tȔ<~ӷtr=&2-[ rz*.&w,syxab-MGm"CƻqJ>eMq<_e4[8fi :cz-|y=~aKyK;fVw/4y6]q!dL|ṛ [$sXJ%o)*h_eT-`" {! &cRٔ&DcCm؎9_EySv;Y7M;.y#|)V{48j6'Mr5>$"#L_Q9̳*>٘|\maLZ"wBc2ê ):a j UY,ؚ\X2 .B慅FP2h췇mPKXX(F,"Y]oa=3#azĀ TP8dNQ1)¨q ,q4q[:]\xIgiɱvQl%E4)&Y@萨EVmJ0D^8D10 eD,.>]|)p48:> SͩKynȽwi-KIpD?~H,0z}|]&=$k^4&(Bl̍#mps SSAKI < G[څːʟ&-ɯݟs,z6S3YN?$S'AMngg>Ȧ4Ar@c\2qrN>  $,"nN+0'e@t$FX1yM̺U\.Zo`ܦr7Ԡ2_wc"'Fs{:uL.CkD7i۟Z+tvv9kVث]2Ti]%brP'#ڞ@ӻuݨ,3/icj3XfDOL7 Nͽuc uRiG|WCirQxNY5j 7FFoMca 3dǯPW_/7__Op)p&ƒHAXnt/ܶkAjko޵0EL9لo/myM?үY[+8r $ՏoP">ǣ8ZN&ia*.G `O^vKʖN!Gt( ]*+w`2B)%xNV1T -\4O>( ؗ6-rIR;_OGlο:AYrnjrk1k +pñGGBG1#/RcC!&MhyΥh!'B3\CsGF"AuauZLYu9;I@MI d\邰.E)뺜Yi`B!s5ϵTd޲5rá&+K8!6&ԫyl#ɯw8=y|@J!f-qɼr6#.gmTdJD ~;1$JE4q~pdǙsX*2NL0Pe|69(!_<NDgL2jKW 4mh7W %L9[{ܥK;f_܌ ii<ׂo %y]6_ Uf `Px@xU*j_Dk3. gjADU/5)iij}֚'?J&ZLnz1*hb|VYv/]ѓ xYnۡWWqw W3ɋizUL%ClIq=^5?_O̪iOuy?r<49vP|C`zo+gȳ[%ǟm 3b<) wD15-Y{Kw!+u(U+&8G "\dZ3cZ(95UӐ 1NF02\/zR.! xv\h;5rVP&G > Z[^.B87՚W7^~ Y>^CND-RN$Ri#pk fa!M0cNj߼lޙ_6)վ^<}bc1MG<'*), /T==MhuѴVdZ"qw>d < h)rzJܪwURR2sC7̦=0cp[4FK]]/+^iU, HL3?5Տ{fVk6Wm7t .n4w{gb5w4=;ozt1K`ΞmMF&v]\67]Z}k&Hrx߮_&ۛNhtfoͺA|dƭ7izo|ݍBhy067YA6<vӮqOhxvZ¥Qo`n[;nMd,;qۏ}҄6&/-knn~ \JUh2kٚIs塺hq6zebc'6_ΊvVK*yb4ïzN6C0($ p`|♋Hs3eHԤ1V@-`iZhYVh1%cdSJhClFJ|>- .c5OqP*Z6LV!PS4LAZ4B#XiT}뼮 NNd|ă$s%,;o*gCd*!t].} B^qH(sRpPJk{ ƊeStYFrIf 24 g pp )q*^bOfq=z<"}iݑ=?rUi A 3cm*6$f6frsJ(-x4m^88o$a ({lK%9_>n< ?"1y"qUp4 *JuqUԽ2ŕ*{D fpCRk뺸*T ً(57fܞT1n2cZR%@i }\d ny9,Ej_9\h֖Of8L$/p[2w$%br@|qI>.lU`1ZNHs2:O O"͆&Fz@Z&Uov":sץϙe2aWV_9𝯫kg~.f g qmj/NlԒI4r$ÝL^+q?ܻWc۳~tEV.xSxfJX3)ɖ]臬b VVvc2'01dZ⥷($y`1iӽ/=vSTbʗȎoJXAg%(|!yx > (DS>IwU{̑О+sd-.Kں_}魤T1)4hwKWv Zw0݉et3Zz(uOՉ^]Ih N{=o=v롕gz(ۖ1Sՠ+K?=OWXR*ᚮUFY* |쒺ʀϾެ\CBWVv(i.4k*֦3ttVꞮ.bTJ(N ]eRt2ZMNW%.PWkM:DWξ<\ޙUFZ2ʶm]J)w݉]e3tBcWJ_{@Қ;DWZsruWUF _ȐT#Eҕr$G^g_Kցj5IVӸ(o@&l'%CD^s@+GH WvfiFMۥZFiT/.P1*ӌ;[o9\ #@dn-IFZ~8tE=yEeYH0>Z^c$iTDW\L:tQlr09!c9Qƀrts Bu&;cf-2J]dh@5PXըh=agԺ!WQAUli{;w}}TN/YaߖW%槚  O-߬?R8=/mɦrt?ZE5oxcUr }1YNV{|g;[Wr3q2[7VMhhljLeu b-#|ַ, Ҫf./M+]ӟ֑OGwlΡ=cW?FvGWc'1ܮ淽~n<|Ub^=7 *5.d\tȵ}Dn6P - ,d+ q2H߫/W$1KWwh [̿|laRܥeol_!P0+շ ۶'J=l,jGyO~|u ( (DpBQ4[42 5 ~{{郶ʯHlo eg}";dPaՎep$x#c^QDC-H7Ts8vnFnW,Q9)1@Cw t Zs&u>e6G1xΒzWdΖۡnab2vniLˬnIɓ@[\\ڂ&E(r'H#"|3D΋rS 7& y~P`)Ŕ}!@>jy-  % I$n)K!EOsCzmx#.*ekJfw9BdА?tPUwgC~;FpUykڿyBW9P^8 &ݍ_7А.D#urJ=.!&ji.j߼3&(/|D5"\~ez}Xu)<[3 BPK3I*!IڴX_1wb.Kڠ)B*ԸE",1)dZ-`0/KnK̻?9.Mw.,5^ <md'Gƅ1$K4 ch)5r1L 8&.x뜡BD D.ͫ (vL5rk [#ܦ4d>|/nG¡Á;e o<]}=!'T^𷻾,~)20PgaHQNWYJNyt?YFW,Wk k53KXl.],J_hzuq~eXŲ^+_.e>Ja2ou𚚧6ǣto#/ln%QՋ(+DL7|Y=g|ki?԰k2uM&6J~cBmIxܧ?aO{=%;d|s- Efp?&2-Kb8w AK+QYAu'LﯿbrCwZU}л,۲Ҥ,FWUP!NJ%h(DN)2ΥH餍I$BzF^8t_?-rLmdY?", fc/r5t+![9$E5A1@ Q7R=I&#hciㄡ>PBDRJh*18UJޏB-FerHR~v=VO[V[OR(4PbBRU"S ^{i 0 \c1ԥf=:UHѧ \hj;08\\FLQ.H0$m18'eKRZd). ؈FAuCW;!HW i8b)(6>Z. ¡J Σ[[stF@KŲ:Y &ȣ./c9}ocm1XNʤ#Q$8G *.IBf4YZU~$Roc-|#A}8L@fxT7t}k}!&+qme"BYpJ΢X$M")5 jfa b)Ei̻e Z4Yt^ќB%4xGgs\Y=[1RYnXm|%Ck YA7#/L:::Dkq ݽ^pB^>JiΟ`}RSR[2'QYl{u&)F!b.H7 N N,Vp@eaADER/.49S>9 *(\qXDWI&ƒR ) i$2t!E̗;EA.G/jGAh4^bIbj@b~nϭ / O< a :QUЊ}/?{ϖHQeRíUY%6rCEHQ Ci"%=yd0_x첱ǧYW3ʻĭ}-ʜs Qm؏:X)FwA+X (\4N٠ >4DV(Cer,֏<AԥhQ 1 EJئB5Yszo^geS,Ru+zL%ntW"`!ugWft%}eF +$O_y+Q{,&GI`Ta%w8 rЃ<>ua/ӺbV[q{p[Aa>Iu"ㅤ &47Jpl :d:>m-XNXp[1*w8W/g9w:o_ (a͝%1 Yc 3p\A`:E.HJph)ބr_^v\.@Q@,@M0“'%#J='${^6;p G.v3fΧ GΎ,qt-};VHHV]TdsqS@$D"}x6d}gy艧Ofw3GcQyJTR,' &%& \ĵW(w2P#>3OSg`~~>08n Y`T䴢+@TNe)`xR1DkM9*[}cV1X7=U4!:(jڳ`~5@6 lBϫt?Gw!`Ǩwm#I"l쇛 WGYI$&)YEKvW?]U]2H1 JB nusyl$U-DeZbM`uf #jŌ>%mdK%ABF׼x0Ȧ -Y_NE!)a}6?.cQ@.aS_KUwe6ȥjDQ5jEP{^YEJ7hljk1&E|Xjɺ$)qf[3k}MWYU UQ^?I?Z4꧜G5EH7a쁀W4ިqPIH7MbU$hb:)B؋UZ\uUb|_lɵ^izIu- ׵ &um+j:+q6rՅ,ЖYd$ϋjK]~]שl(5RVl&7$Rڼu`)%nTiם5 gߝSbþ# E\R,C6&W[crU>x>ئ>dI٬ݸ&-jv#Ul7T,}!l4]JIZ"G!5j7Rߙ:hs%vKeh=,u{u(v‹0zm6noT]iN˖8NO\op7Cɯ'Șo9X 2Qu:6?{K{TxdHn2hS3@]ꭠG{|[y)&j^4cs׶hbgiףM pbN}Ϩp(h#y >PKG.p6*PRd.(!;J|-EH Y8F|bQr ^cȐ@)k+ qTHxJBZcQ!PKE>ٯ5lkq7fk}*UQ`t_-ꅝ.gQ۳3R2I>OÇK=]^i_n| L$Bv NR\DRMkd9"$UCve;Ia~&Va<()<L#{B %°VV^ x$IsRXCaj3jXTOe4zl5-]D[)K>ؾ}.i`P`ǤQZ8 @3$ $D8Ģs:,qE|?a ߀F.00L/+Ű[Pko[p_8gY:O1|;VjKr;bN騎J~ 2(NCM>zVTXhC^KVYehsV+ʜK8bXJbG5"ت[eFd>Z7  N9ťoW12# Ɯ$63qflOWi [=z5. wY@,f\,/˸f7sWU>< _ǣOL& ƤH;JðD"L4H\$[OguXL`ͤwnv:4&h$6d|Y1,A!f`E;$p r"-Bx{Ҵ]ץ58aa/,1v[}̈d`āI!`r"6lk*B5PdJY˄J QUF #$d*QHLd$&:R=.,iPlICFbֱ583/Η]u&%"m/nxE](EQ:0"` `q`H/kΔtcbW5ؗY|#P#/E/z{^8 (+>7D?>Q#E|^[я˥Qh}zc^;48=$F1cr%)l,0H:ծzv|s2_^Ϋk눁Tġ8I ˥CRp1AqOrrQ EԤdFk:n+SHqG1m :;;]\^S9BR CFwAk8LU!G}s(}01LYy{rEKymqsb Rߎ6s9Tߎs$L'?]'\w&B5=)k{b|sOuݐnD}7iO0eT㧓r}.Gm'{%h{ˇY%z'{X"3$ϏәKM_C6YS|i}χFoecIy9s :oҿ^_~z'Lԧ8kq#6:eۃ r뿟Դkhko5UAטՀ76wU~ cvV[n(t0w<_,N<<],QR]O|a_r:pURr]B nK%<&vq/gn#)AIOKT\#lќX##| `k)&>xεllXR^^uIT`5`8: zřg b#1!RSk3Vt:b53b3e{"{dzAv&lI^k5lͼg78|n#MY0KX͙" 92nZ(Ҕ'&n$FipҔN)3aQEO<% Ĺ(E#͊}^R}g}ԮݸhX ိ`fĠhq{v ^v{[pW+^R.=Sb \X{{¾\f|O~.J]PZ(1D xIvsSʽ I>`U#4VrfS{%R+:P[Avhcrw<޺Sr7ϾI;T,KX j6v=lb%K;ޱ>.mڠCQG"`"RSF@%Z4`%טtvw-uSywufVfr)hwϷ>6ldmȱZDrca^*jb w MX]>eۦNxQ 9$=KmGmm `< Lʄܖ=<10c~ n:+j#Տ+u/_(_MzJ5b_s>9 $Gۗ),W8pZIv*,;^a@&dEzqΊm eYxY)Ei{-$ ڦZzWoܰK54 *}_;fV6~V]y_$XvլXM%AﲲuW2~*IҘLg+1JS-0]%?t*UBI@WϐVB7t؃{8]%k:@Wχ‚H\BW -1PJ63+}Ү$bUKzJhy絫r=3@Wφ\I\Jh5:]Jv<4+(4FtU] Ђ2uJ(ɠ]=KX[ee-b= r1Sr^e97UUOۯ/tQa4093"ϰ<_{n D4`>]RDt7|v[4vi5SJzvNW eND.tER%#UKX_*etP 53+J`E*V7tpwp1DW -C]!]1R8CWX Sp솖uL(zt)T9FJh NW %HW~iWXhW . ]%BtJ%zt% ?'VcuJ(jt\@7tZyc0dC s+M g᱀3J #@*rc5WfVioh:/4Њp$ 4iZrz1if*Ѧ3 TQUNƒˌ8,ŭ˯rr-c1 E\8y#s1#wE,PZU䰿Ƙcbv5=[s繚8 @|f P3yMìwO>ӫTEl~3)d=/|*,v"?)[H)np09sH[G2beP2aFEF٤VAөla:X v5 &uD9hkcN&zG̬X 3QfN]^r $gޕ6rkٿ"LI,o2@7 |xXeɑdwsoi+ے,eUI:$/a"ÜL '7сDa?`),f?JWbIbܢ!m'mz*VEt %+1_DWlچބBb5tpl ]!ZNWҲNښ6=5-th퇒v)ҕP6]!`[CWW۶bQ6-eGWoBWRoߧ.B5SVT+Q"BʴJ+u%Nfؒ,d@M+Di::E2ZK z< *BsW ҕLW4i*Ps(.;[DwEwg~Mvt 5vKQ_!ߝm*] ̴Z'-! Xeb y=MzO=4Kv0=)'x ŗIwa"@[S3\@X\*ϙR`hoegS#~?EG*VS5JՊ"ٹԂs*vs6N?A=3݌9*4fG]6hgۣϺnۿy>>@(Y\5lLk\5(mhE'eZt۸j2԰MK/8SL1i#>Vy4𠠽 H嘽fG岺;q%ggƝ4mVKfkNLY%lFO0z阵чלv@='z>v|=H9nPچ,=fh7phSJOWذµ-th͑‘wtu:tM-+,mk *BSg_JWبµ+Nat(+AEt5++Z3wh5k:]!n$J׏_+]IɱϦn{ *tBFvtut䜷 2ph ]!ZNNWRN4ӒiA[3whl:]JETGW'HWYc$mїA[ѾӕTvutu:te9׌6l*գRK.έ~1qK>U 8i FNd[hb@M G h*#HF7>}}+~֛5gOv3ETOWS<N)8@o-P]]ˢVG xJ>fNT:y6QB#nhu 47:5Ӷ1h51no"{-^w =X{Ϸ{n4\-^w7~,?R' C)DrBQ($-YY& 韀dklb D/S!,_(hƝ?)Iᦛ7lM~W),=%e6hM^ 7Aq*?~ߦ WI-%[WڜZ˜¸_./.ӞF3yKhJC %,u,+sFjwUkg֤eNj QS ,M(b>3"h%TInSsu$VfSd2nI·[ɲkʩ|,Z\Ǭu汰ԿK3pY .vLp>'ƅ)f4YC"(Y \=ru1NɥdfoM)zAyoLb:SK1̺Jg y~1H9e4&-1.暎R ^Wcow(djornhK~Ҫ̻<(AgL9-z=YX)P("W6,a\vze.\j9U2uNWgLllyf!c̟{4S[^\ܻaw7J9;r*NI[.L|Æ.ǣAq3pThaVJPA '5L#2©$ \('NrܸNzW7Y>,~`z?K9ga*내4<6l TŲfilf(Nؐ0Daժޙ`G`(n`Z,R\Qw2)93@7(xSv /o6hZӻ n? Q{e#YzyhLL(cI[Ǫe?Ns,^Й۠fb~=Z 'LFeP u7ihWJqaڵm^l^vT舋F_rac \[WYNU}9_e²ymyN IWA Ʉ$Lda|b#)1O` C"&ČImf̺h:m4DJrL"ixZdH"5{yK]x7A?WA9\tVAI$(B##lw o-EUi#b>8.5PUS$JMAZ4\Xu^EG\J(hȅV %!yeR(2wG @%QWjS) )>爃3q :Λh{ 5`DA@3=MxyBRoP!MGI eNj`V'ϼ),eRtYF,00t}I}LF2+)$0Sd0,b 6 nPH䮏5Sɵz~k'T-2Mi:iV9G LqN}R8hx$[1ʫN f /#γ`$o)W7d #g ,ZKКiCF' Qd&lUދZv[^J KT,F@̼i uܭ cj&}EZby-˙'PW[1s$G0Dƶ&&Մ%&4eN$iRG Oc5+m&܃`f].s9&!-j}ЧE"?iu7I`0nY` # CE1{Zq6~@%w2 u .}I*PY8#tN$ErRh%݅_j|%% f +%$W\LL,T"DdbD`}0t񀡿 LPNo L2)pr'%w=ZuDyfIF=ͧGbJƹv G1Okp?Q&TXgX!5^DVйٹ u?V)xs (J\RG!)hOa&,Fb-1M?䋎2c܌wsO]d]WCX,NQ ^4]𐤣Ԡ.Ҹ(1R#K1SO+51h)IGcRn˴Tfo#:9dgnNj:Uս-.8~Џ1h޴~ySbO`uE,,QC_96eF,?td3 t]oGWÎuuI..ЏjgTH*pUσ$ISMvTVUNrbIDď~WGڃ##HH;hd_s~ya[,&4J-$Ņx'@%cԻT:7^a(}~~}&oi(* f3Y^*Yۛڞ_1p1q(QI)"Jh>WVFsC >o;x_gyOwsWeԯv#m 0-J9;=yۊ9nq]:d 6%~‰Dg!(8-Fیl'Tٕ p݊z  [oBlh Aw׽ClkԽ(-Vn"7"n}+[}aM߰V> .$`c˽fOlǙ}#J(NeϸVUSI9ǹ|p sdݱ5*vk6ózn 5?Z4#əHk[y=#Pzj{V DT@|b26»k= ʈXv_ͷ9`:uNdq_;~wdMJ%$‡jҀ٠ HVڗ#"=H6x_L6Jh)z gfv3elf=tUS[Gt9`w Gcqxxa̻I;Xdjo3Kai@f8:4 =F0&\DILȺ@DIE\)PM**-S %Ϳ!@ty +j1v6+*ԡRWľ9/Rd%q[Š29OCCn<`G=Y'=?D碨-LfU!xo4,jc$( TkӐ8D.rKuQdTFHh? &cEѤ$PR U*Űd숅K`G,HE :[=ۜ4vv3 mՓPhq4ͿpĶ2yt&}$6Bx^;A5qP:8R<@8ire" ;{.2½geJ%e6ꄾ161aAX$)|;Oq<.&6-G vmJ!0O]WB࢓l>X1g5"YO $4kaJ-C*Ћ4bMfJOd$jЮt-blک,20 "]FDvD#"xLcJ#>;lڀ&F8,Eo $ QD+PBQpPt슇0p!)Iȍ=6aMAp'~Q~lw &*@j& ^a2雄6Ǻ~ /_"-/_ou©G?dz8;?MP? 8v5<[KmQ\r sԋ3E8rVW:k=́;,A_o⬉/|6? y\ xm <쿷YߛI^L_xmae pE:UXoA3YWڎ@;xրAy=u]]dlqp3.XN~G\ZS?^^Ń)qzϪhSUP \eG;e澫1`f#woC`ޑpRL&S=;j"T`m;yU N\53ꀀ}6z. .eD] 3x%LZF5,cgˠP5g--~apIcj{ysO"}!,6~oΑy-wFt׸,fgKXbFd)E(*5"dT \ObJ3Ktb$$AF׳|o.6W10i-y BEI*H єjpڜ{}>XyY>uȼIs<*xЊG`F%!1E/U\2d~'8s-M(RBD.' :r]w=Ղ+%sjgRTVv9qHy0p|D@\Hd"(/U*ẕAg"y*W+h`(m &!ęNOߚn%^V't.i&s6o&?v~PJT뚲pNĻzPCXE,oވj8+r'Ny8q\lxy#[׺Fw$Pri!cK6]ʈe7 y}2~_aJջ0q0Xo~A~m/j:;;_^.Tqb%!ZBwy4ui׏6߮qphw{ܞOzʛ«Ee!3i{;ϗ8/vn6G Q;KAH%i/%!WS#ɨ~26xR`S6xtMr8_p*#vGY5,1!t:RW Fn2 y &ia㔞9Guχ.A5 A^]8 ߼6?ǛW߾=̜o^'9\NXh?_ŀ`鿽q[ҜU4Hs# Meɧn'| |߯ktkgvKn@xǫqf8 9 hU=a62\ v_AQ/l!߮RW2bػ|q[ѻ>yXnNzvg5I>oC!HH :2TZQzŬd>dń'U\cN(Gso6tWCp8,ƨ%#@@:hEqt)%' 7jLșAU,&;[!ٽϮ';np{WU(ç2T+V(*ec_z1"p=߂6*b"AdTSVƗSVv}@:~E5=cr L&[_]bjk|ixC.CǣuD֑Q U99MpǫELDDB*t.pr-aGlT{5՜Qsڑ@ Q'"RD3)aSIsF; ec6R!:)h`E4U謏 } F9ᔕ@TNAEo3Fg..A6.gG ڣ 4. = l!žK٣-^(s)̱WX+E-\&u.wY/ ț#/oTľI`ZWr{\8 9䉟v߻v;'Hr 5669t2 wmm$;rvhXWVyfn9L<A]m!x|ߗպXՖ"b9Qb[bɯH|{; F~LJzAzm$Lh9f򽛸2`'Yr RX" JgTZ"/B%^_ozZیf1ۇݜ8Ij2]OΧּKI|NRƸRs^P})fU:dv0,2J6% T': UcN᜘*VS5v6kZ; hN /XS68)cy"2=m O# Hȯԏ'ʞY(,l"m#ߊ?M[4j n 2~/Lq:F9=Jt4ۑC mfl;kD"l7?SY"#0]ݶGw7]^ lmr'!lv Z]wIEQ'q:Z| nVE+zچ;[mH{Kyzu]-)l.Y;8-B9󷐨م޴Mz.2/fJw0Ҿr&ױ]ђFK|>nߣ?ҦbP!G N߿c?[qߎ'"~v:\{211c2[N}׬:6ozt.) n3Kao|81A6\btONyg94 [ʼnG;ͳ[,;uK.uaF #5F2Rg/<|;nTuF4c9F u0ѧRK.wHFZsyuT Y,7y96*ls &!E71Z 2ǓLap6񠃋KdQ9lq 9FdcFrɧ/h*1r\ߓ!yJͬ+bWMVGҘ^18-H rj g'7y5vb'$-R&2ؕ^l_ibdǷ0+<~74n0 gFlK.'gGf2&"x$äMAcX:Il-PdT=9s l m^6LCIZC,ٌ~4K.桠v52j죧{\WYdE\`Ν51s\4U0+m\TPe !Ǡғ'|*@Nu2vYjl֩'`<D&""GD|L 3Ӝ PfP8l l N-F)N5chӦ=)6k"s}@Yڡ K#IZdeDfD9O=x~bդdW\qQq񈋏I{}>E{iz0K>n#CJ $ɨOFiQg"w+#l:1IY|; Gi\K6,Q̅@ITڑv`$Cw@j\r,XR>*R]E%h\r B&SV`d^qĪ:э:z茖\uw.|(p^,WdQrg+-8?oO;C 52̫m{қz8΃2h|AF?zAMvog\b>rtU#P2D`䒝$@7 ?=O/N6k;$U@3X1Y `Z.p&؍4o Yi8i~_F]u4C7mtC]D;^:\\.fcF3UZ !Wj<>g%xGJ^:q^2JޯVzҼ<Z?+?^_OgY62w|v>[av)8^![L u:GjH]at06'iIjG|wc/tkԮg5 u\i#m^GW( l F P<$I&9%N/BoGx?߾~?}߿7V LREᤋi' <7n;`Lƛ-Pn14O=-#w5b֙Œ*z^zCػ/;m=ŪIq\_!I/ ^xPںJkk\BweZv}nd2cKZEl:+bgf՚$ ۬|>Z`"Zh42[}(Fhi%'>E?H;0՞b󌜃eI8\ ~r0&JFI8  ) ItTEj2c\2v9n߹%gYuLyŤ#Uv/eQu݆FI#SIXHTk$dc5J[W([ӝRvIYhZʺ/ s 0Ie61pĝOFr-xkfE 2d*.Sv}@:^UGzƔԳNctr8=we̹ͬ7Zu1@9qq[$OQ\HrP(V?l$4 GW7eU{#ơ49iLu6{4c{://{6(@bWE5OB0g*sB}R\aJB K E?{Fr~".[~:ˤ"j篚/QFHZ g8S]UzJ`+UKd{*:딮C7k ѯ qd0ږ>ykl^߃"ޓEb5n@)t_'h W.0&QkyQTqL35A`Wkl( 2&B$ Ɂc@&Ǥ ,T%[]H89;,!u{`u"g*q܍l~ؙ'p_U4Ei.݅h뷧 y>/ 4uOLR5K:;քq-B1uBU|@̫_p:Wtwc/["z0>cd[a E!eX2F4 PMN]A)<߮Ifj]JhGf?&U(dv/~w_E]uTYP&͘ \H]K.x-:D([r*ɆB!`QljpZg-̍àfpTM9+Ue-;<2pvNPњw#JNk i蠆َWD'^L,9ϷЭzA\85{ߎ͞$Hg^V3ƶ8?拨S)Ʈb;C]jMa`F;lQxSȘ<.o]^!e'+09;d垭tT{TODu̦geՓ-vDp>T&@WjGIqMdXr1atŏꇪ=c}at 8+Kΰ/IqzeS#gEdV0Tu G5L9@\kyx+Ȼ.~Ӽq=dݵ뇂ӶvG'g1?Glp]ʵ.cǦ*Gb[;bZރCb]Ug낚,\9XWVU,Y[렧P5#9 `U9k}3 ցM.tu0rv I^&FG. hǛ`-ræժ^ez.˧kfvwtNf[#xl(wvrg{~䫛?|Yn'oy7?\Lw>uB" ɿils|l#h X6j=JcX>:ns8pܾ[>m݂j9Q \Eť;|8<« ;mdJAi@H|,XQl*Yybɤ|I= &UVI9P(mpGbHTͥx $Ψo Gފ=YZtr51/{W Ur%> aYxlK=dX[G8|ѱ&P@ g϶k 02K.K1MyJ!UGzle8Hpu%bZe> WC%j2lâr5gnmF;$3TK"8J$jHU6v,|DuZ"?eENr}hW O/tzz$|{z~f]oʖ&.LĶCuGA.Lp%mڐQ\mwdÇU۪xA^kБ@j<4PHuA0T@1BiJ7!=3΁w3b`|8ڡ8j78D×$~la,aow~&sYl; c-h|V }pe{@ZQhGhS g*h(`EJ춢)lňOV} Q ~?QV;xV﷛{ جnα{a(ˎ)E5ĤbTWJ!J] h'MNQ)Fng/2u-]vpa,-|{w!ȫVykk^YnvhvJ9:!6Ž=頑t,IB->u VI/0[m="uǣ^jԲ?tu%TZQ]@uź߀!井dnnL޿;# EwJN?>^.n2; ָ *m|1 NѨFcQӍZF5մ3 @ogs~2 _&*g6od9'Ar2!!>_l Ew./soKob'>}<=?ߖɫ9_-ѐeQ?Tcbݥt/|>e ʨ(;(׵r ΥG1$z&}߱G^0̘śhGDVk6ђ:u(8ڵPsB"(6%6k, ȱO±:v^7Yȼ,M]-~=WjQmx0hb?t!1F30RWRt,QtQͨ^B'H] G]5rhUUQ]@us8:[.;9ƖrqNe8{*sfK5R"AinfcغV(UL[;+o#kܱ ;Xo F$[f#jju,QC߈ƍEnD /27iU׍)/7ybs?m]^Vդw|TUT-4ɈJC$)0x2bJN޽~5^jyۼ~XO>pPE}h>W(?vZOPNtDs v?CruH>+ えB.٣B?xתPʍBDR7.ƾ!cߐo7d2 +}Cƾ!}Cƞcߐo7d2 }C]L߳GP$z{f}5eh,uOʬ^ˋo;([8Y/<-4ןr6TV&'v%brAwκl$ZlD2OoHUyzL<$aoY{}}NkQ>ŝ7_%kb6[,Fs> +VjCΦח'zh2="'T=10-=rx# a\ܮ<5?ٖ\EC-M" fZ)h"8N` cbۉASlz?}\&?̒nP{Qޟds#Ch?k;uE.5@RT|N&"1;-E e4aTٰ77A9:"/ZeZRHZmhЈU3բ/kCzځl0e퓽ܷRKA';$Pp(z.%*阓,Qadvtclum/2^3k/#OMRRƤQ[WP2VSd0: 2RYdRXK \vm$dkDe+f$27:&5GݺLPbʺl$lhf&&/{ Ck NnEmڎ( pf!*K%F2-$8}LEZk&.2fkGKmvl EGe֕1ߘ?,r};`2ZFZ'ʚpy-1c$Nhy}7>kywK4JT,B=l%)@-PUkEUPA˶.?ʦdEuI>jYLK͐qxw+"O`A/& !jeNDrCТ1 )r/JL3 *;*v &Evʾ͢a drUKX Y&єCK.ޢ)AtP[>p*Z"^0U>.$V 5^ K6W6yQ22-S5Jsj͛l}Ċ !ڠ!.Zi(4p,sL-uIiaJ]/8nCke7̹gZ8_ G&$^}览hz{n @0 E mRj"{{oEF{C/F@ri+w%A`K]#TA4d,4pvLc+ ~ :ΒB\cb*2%h3 6 4 Ao`Ù1?rAB}a<*#O)è|Ձci`2h,q,> vy9<+f@57.WR̈Z c<o 1c;FڝbQ' s9_.J)y p`R@gZ9 ` lX "2I_~N UL[e@SѝQ3Dwi7h,X U,eCcb"<WG50~S.pP*Z!uϮLD(AץĨ}J">/)@g5"I̼X\jEQ_U0Vr!bLc`c DD[0,pZxZ댡"aȐԪ_3 ú1.YfV̪NcJA;L'A<waZ6N$;#Wir¬̷Z.xjka&Q:B%W4UБV%^E H 2Jh ɠ< ;l _jTcA#ƅ1 $bDL W$2+BC,Ѹ~{ ȋ ":t@2Q#&@܊`mm}ƂQ,TG''zO(a0l dNI0HNH6Vx`}߭<* 7Db6y.e}t@QR"d2=bxG:&oCUQanja8 Lc< hwXEy=*f@S.-ȊAmJx -Gh0K;l,IaX0FaE^Y`Áq !dg֜ÿ?peVHGYtg!LkTRp0xAmJUΪrM2+{уEXY oT[ 38]$y͸tSP1z1sɚBac-4+vNiWޮ5j k.m'ԵEW@7]px3 \U  -~)h;5,:ZtkH޵繨)UFCon0zlihG3wI xXRcF4Jtcls m娓Q.1eBw GRDQZ RЃ1V k&ix z_a+@1\ƒr\ \1DNa~;?" <2apaB7B-JFJQDF|RՒJzi,:a1#d&hâHY9zV"5R&0R7kԨT@j(F 8XH#mIҸh ܵ5#ΏjyAKzp*R3QWqӆ D,:rfF8ae%3=f@D2z/0#ASd> LPb.'ܭ_wx i0?li3 9$bUrd"&ƲK1!:Fff%$UI8 $‡0@AI56B,?`EoW$FXx3= P GM%B?e.n\,ζv@ qSbheO?m?"me]L"txѹ`D /]Z'gٴ疉;ͬ66]iRNOF;2/o'`Q,6e0Z /FƬ]H3u21+whBј-Eu^= ͜gNbqVWGgv~}*mPebSr{xEӘ4&(414&DӘhMciL414&DӘhMciL414&DӘhMciL414&DӘhMciL414&DӘhMciL414&DӘh߮Ic!8: kL'McԘW_Tc7qEdJՎ#u. .oߨw.V7ӳɫy6hGgebhOe44!U8C UUT+^+(yYZCQ5P}~o|~̦qVg8eks /il}wE{Cŀ_;EL2.jfw%ݐ}.J㲰tYsTL<6p'f#ǠYs ߊc X`} ]RVJKhedW#qzc8c͜<_?Ugz߁37C}"|m.g.4Wxt Yd!k`Tj;ɅixI÷%xqj 5+.G l|igw8=3n\ǃ ɛ@jW[jQ/TՕ\Jw9_y@O e"_mn2}9.qyѷNICWZ ]hkRl g/f+s7˓W_=GC9{m2Eׁ>Hɿ2~$bLtvs8zN^NAs't`EiIqX_O? ޚ۷;xu8LY* ƳwG].B0 7wU?7p^U LG*avv_٫88Onz~d>çϾ'ٻMxe/~xq)kn-_ܾUOoz\`sx_Qѯ/`k]O}Cg_y.i<=HI}ãGAJ+O&?_]Erճ>4">՜gn 7_漯7(aj}ɓ>$i )mZfu Oi}x)lI&iH>oOfFMwCkRqyn6tdct Mجncf=q[i5m^=|GmNc8wyA}i MMVaV}>: \|=vs;㮕o֒c,O.oz<7r ("gc=~QR +CWt!TJ#T[Fmꞥ}/s܇s+3=8^S+I؜=\9ς(R2xIH&,1Zdh[W_F[qml\Yqe7~_cM'%κ{_uKp?"JTSV)|lo[Vmo[Vmo[Vmo[Vmo[Vmo[Vmo[Vmo[Vmo[Vmo[Vmo[Vmo[o-"&*|][o#7+B^,ζ@ I'W;XxSպXݒ,S43c[d5,~UJgL[wREi·(f{,|){ ^4&3jn5O l5iT5VX9a VW-$Q%ڭd)_Q0mK쐀28 `ԧ%VS e+^"F`{1q6Re.|wf/pT, B;ü7҂f;PF7I3 S"*J΁jZt)P~油+:]#kn* ͓m*U wWË.U2%[|}aN[`JdkږZ.ljw0#=~(jj *6o=jc$(Bk3Ǥ xjKfԌYarlfƖpZ.ܶfՅԝowM.3fGw0rC$'+L&ƤCk'&FvD)89$ rjlnxs= &W:`<Qg툱 U65v1qkl?s_vٱ֦6Z"ؕ5Ѭy żw#Ycpl>d)cDd9yj*,vIa J@T`0(OtML$( !9~p"3/ekw(-ks>698l2gf̷o\}2RnbRLo7:xbm&`PTyg+*HH`YDC`c|ѓ!\:[2~,o~(Eh`KR4. "L7ɀώ;^R6ypduE(X~*5"lT \{)g\&&֭XBvB=K4Ay u.,M?4%-ET'x4)Y"QYHb sKcH.!9NF?:_kkd!e%`;I€1M9c#Mҵ AJɜ#$3LaV˔B*aDr/`@QD%Ayzܨ"(9N+4#rz;=WޅOhͷW.?ƽ&nl#щo;] .FOf9Ͻs7&?Kw HHmJC{'NXPS%G z=B9aGW B)Gn4Jw4us½>QlL菽gݛ܎kn>,(8_/ܿ8;-c n-~ = tO]ݰnTw7gX1 ~Q]|8ztx^Y!zzV,'^GHfkdrbuoF7=TF~肢xܨ9뤱r/\\t?ߟ;?>̜~ӟv?SMK^,¿^$/QӦ]sBTY܈ SO|~/~ͼ֘-]@tA( ݌=Os]sddN|Će~99´V?J)Rت>hGA{ΓǶ."9|JhM*|+i!fqm^ Sg&gI4 icIP9=hRqnp{3LIgJ8L2D Ȧ=)X0ű`E3IpI1Pg -3yD/d;t9ό iy.\^  |kC ~PT2G3SUkM.ϛg!ѥ*[AtxXh R&m uU VaRjxs 2 (}\NydNNT n|9-;:Ky ;XhRlS ,\b|# iMrygm|]&`v))@[M1aK;]z)O~|w_A%5"9x\(a~JVj䢷M.N{u(U;/M)1[rHA$BKʜ)՜GC:)H# ,Qk$bΔBgS\VԅA YW@'^,aۤ\hAn3~z/޲*J\BK7od߀Q tWR* A4~ Pk'Z& 5L 'u >`;Rk2d;om: Rq1 Nt$IH/DB,~Rg0Ihȍ6^z  @:*{Y09՞#i2&V 6)UMwJOc*=h. ԎX ĸP4l&kHdE2KeTEAtJ.%%33xCmJ } {KHeX=RYL)&Ά(㯀f0k.:  .1pqubĠv3fo#xc η8]6(zOmx1f4=݅7ozx5n C:+Eav[rhn[dFG  Mnbpǻݫ4:guN¯ͶA;lݼU7>5V6hyz8\j3o2moV?qOi"ƃyzEMԜ=MkuMy~ %HƉG[u%+Q f: W" lA\KQZNݕR ]]Еhd\9P\FE]BB)-+˅t.p>mTnNzd$g/rP7MO_>w=On0;#b5?0uGFMCQ(miR5-6|HJ|su'*wv=1;^0-1Rݮ4(&7;*e%9-1E;gRQ]){ku|?..`4P ocFwo Ko ~{7˩P{Kmnk7:q8 'Ï0#4;[͚qm.k&Ͽ}(g]_&]l?epxӃqm w|L&9舋7Frcc.o}a.S8 w%ۣʁP ۿ*(!$89,O,3t<0c!BbƤf%mt@4 $.*- KC1gZȍoKi$G];M41Er9-5QF(Hlz&ACwAUp8=1BIt10ZxWR1rVJyq5{pAbƷ&7Kg˨Bdj 4, #g|7UlKa" q&M0V}rB@;"E4 4T6%B4TmxαFSx@ v(uoΙMS)tr.YK2P+hXƨm1X^ʤ #STIpVG@0ajqI2mVL#v2E26ogiYE2}6m+o7y׾Iq}pjΪp9tF ?\AK'+ሯ";fsԺҝ^ș5ܑ;>}vE?]?O;+>XY>cJLF}76hՠ.%ʏ r*J0TIRScwQzKW{>]أnHKyѹu m<JrŒhcDpBViÂ+#2<P܀@5U&ƇrT*FjA{8ka9ks4^\n|6wp>9d܃j/Bխy.^vA⽲ФczRkw^GW4iMY-iZ6Kcglׯd*&}*u9Jv:;;;-L`0y@ݼ `cDxNcP.T:/L4hl IUVibIs4A(.=T2FKnvv.o&o`f.[;~~3?k#|aDI <&z\t$ >HJ()"< "d^,` 08Je ,#F䁢tD)4pmB'yx>܎%~Q5 Ors-E3vS:xn' N Ԑ\@}Tꂄ: HD+9rdtx<4젧zv!Q%*)BȥƦ$yN.hZ>퐧Ԃ'2@?1rp|e 3 VԂHʋ 9 `O*FTR5娉؝5BuբEf?J(YZEԽҖ]ϯ&p^t}" K%@< 즛MYoKnf2z7Gާ{OstKSi]`5Wî[XxxKp6OCYB|b&MU=P??/׳Ğ_.^o[q}0|?:us&G>a$$VTTmʠSQhr'Nc\!Z%A$俟 G?'uٹ|i1''CokhI޺I4'>Nsy/gKmΑ8{#''c㏧toi4.%}]\{y1{&?8R׽E8m8{gw96wxJ/np5C8z;v9y&IVwUtiKCQ!/Bȩ&ꗤ6є~;/gGQ( up^ )v~|.^ˁ~{BL䚙D}i^☾QEg؛M,>o'ivOuQ?2M[yf^.A\^:6Wy{9ƽ34ݦ$.Z`efwfE=a=᝵'z&ns]\u4qo)JY23%;̦ޙhcS/i  T(ǨK8JHm=%hbbe),,*}{QA ,;Azi%ZCSa,q,rN kƀ+ .p& 1k"O:Vˀ^@H'U'_~_wv1޸&kohlo ;Pm^XiM;I_k74>RHyF3]\LBҙJ=:Dm-S g}|0\27Q PIkRI%bՠ++4OLinvY,Rn]m9EVddP &M%s% /{ '1)Z WeG%ˆryo2ՉKYh&2U>/FΚ{mrYib 8bB0MymȘp9*5ȲH6 jZTP 2b % !)2T a(< ΛsRe,(,Y- thEq^uM8 ɦN5%Nd|9ڠdzA4l&W4-5z-77?l EQyEW&\ I4H=q )?8 T])=1$`B':$Yp11(GG.!HFbQR qƆX\Yn{[lMγfܲ*36W_0?rcWOqĶh2y4&}  !?Դ %TK$q8Da"8o@JW@cd(v)))jN &ErG)+$"&ɵ K&@yz"ިI: N+qcI~XjO=]39+d3#񿺝I/lؘYr~u=X>hu%\ԷtуMGMN|cgir^9AȾB_C;S'9GmY۩''}jlPN)}W/ ~jvmiZG*!ѣd6Vr)#pփwrp.sPLԭT/>gp>6ީ3r 4w4OY{ u+sBe"s}X.rGSٍxqQ2~wFgcBw=ZEʫ‹bvŴ3w~t:^s3~> >89D[ n?$ipړt kFw,G1 |NY|8_w֛pWFum=fEuǹ%Ϗ1NO;éc߃_o67U6L/w7:Evѻ?z=zo_{ez3p&pRo#˃ßJuͻF5%z6kxGn-E1ۃŔǝ/^ r|}Y -W&ZĽpeNΉ zufM||FR޵k_6MAwiL_I05"KjIvgTٺdUEՍ8NXd-In@CA!~lK e f~f]M#G҆xW)}P(Iju>;G4$wRLCskc¸Gz`ÜP|ciҔ#98z-Q-x)#VyRB3mbbֵs:*#de}L⌐T)ԙگߡjgx6Ȫve-|μQ HhW#!>MSx"rZ#α@.jk#M^aV%G|ѫVfgɋI UzCNk%lF0lBK%ծ}Y˴&c&bbS;c:OZ#QRӮ?UٯRN٭}qfLNA pft㞟2[/5[ЋnUGRG (WM4G ]~!fQiq' b3 Lˆ1FF  l ٦V/-r5.O/Ť7< ?7vŧ>nvɘϠI :MZdi6`џ7 ?PUl3b8)?"@VsQJ̔U2δ$,gqמjMAI#IY:%U$%c颠Hʱ6i2&g"\MD&$S#FkuN,U}]epအC)rmG ny5oT=~kN~:'-3 z4_Vb^HbJB5|&`)^x>lP}Sk= Jy"nF82(ɌlgMg$HìD#AKmd"5 U1 CH 0)ǼVi.w_TN)1C"0J'6(syl՞S ig>p4`nu4[m .e7EK@2 /U,ϕoY?{<i~Žƾ_+fbo_z,AgJ5 T.Dru­=j*JI+v玟fӮޢ/]GQIAN".Uh m</Aj2;hrRŞh,5x}) x_c!j*329 {p 6}tpFhd"2q[.XK˭4$KU纯 f\O<s"Ǖ< r{w+ <.4gGh>^w^u \Ĵkø8mqb|^e1rH.Aۺ0qa$!m$_kۍW_8|RġCbwkj J1*@3*wMSz8q<٦S 7r@> ɲ )(|+oVV꽷4!2xFkIp#41ۗTo?Hf6O|ۯ3aԒ[eZSk}7o@22'G䞣xԯ"(Щ 9h3Q#ψPnE9$YT͊fڛk#םkB×~`mJ[Iu!I8 q.Khi8O >SJȵE3 8:u~p4-l$o,ND#ng=ǧ9.KYq&{.YJnv}ayYx2{WJ9yv{e!v22{ +8[R/󢇞JeeIyh۲[.Kk ][_7W;Sd1R)!"6R+nDWXpn ]eBԝ2JZBJDW0㍡ 7F]eם2JI[B2LHLBKٕQz|v~Yq~ wVCpb Fy'c]:Ο3NgE& ,Po#qe\c!nDƆhW,_|m˞в%\|Us E=eɼ 11A XUd!)ܳL,dڴ& lzq5tu`m.LW5tuZCe8+ǞՂt%Z:^NW.MVUFraKWCWT+1t62ZANW-]]!]1shEtrĖ62\šBW-uR57L( 72\uQw▮4I] #1o3UFx*Ժ++iAt \ItS*=Xz(%o Jc,M +CWlBVUFtu=te0SL72`UFj!䭺JbXTAP%K 1X MܺUXoyإ/<#zZ-/CGfS -MkY&1tUF+U*T++J`At%i]eL7"tQ2#&+Li ]e^ovZCNW'&+@is% ~Įд++AոAtjpycUFjOWi \ fZ72K'9 k+E ѺAt)Ǎ W4&Ԟڇ%k2\%]iF +qsB.5MV(UjF( ʙ3(i ]eoj2ZNWENf%^TvR GEcih++wxQ+Ik~6qҢ~I(jd{ Oc9ojfsPQ]6cYbҞ?jeIʵ>4!Ǭ,Uo O^F!YPУ㏢l4S77}6- Ez?#/9}A}a[ඛoY p|,jB}8?߽ڻna|4f _kz;~ 5Z}6e]ɼ6[]flo4{ur>p4-lSH{ m2thdq֒Mg`el>c@PA4m`vSv|O0%Deo=~e0+%wpvh)y;Yc3waD>Q0?~M믉}]s^yƗfP{ 4Н]]*_VfZlg392;$6Qo+ϒ <>iF(ͯ vp>7W*X6"$<41ᄚ'(L{VO"Yw;z)T1>a*KN06h~F-j2 T`+礩ux}> Nٱ% dGYSZ p"135)g+յVÓRt Ay✲Фc:Zk#4X#yۖ QZ$|BD @FМIT$Δ:vT(%ST1AL7ZdЏ4lrݐ.;k;DN 0jk| 0R˾X)'8Ĵ"E0?QTY0֚Q'*BBí3?;3?ZeR,Pj9" FQK[D/R2ȣ'FN'V/~ O.\Ƈ 泓札\۩z7 _of$R  xF+X (\NA>*nL8#F(kBP5J,JA]|7EZ3Br@J@DRlkGǹ,UZʹd'!ób%AUM!Ol-߭N:uFHN+j(ʉ@\q_:oA'c\k7J_+ebnk01I[Emp(BJxV//ڕ93(i^m8" !oLhZ8!E2&g8y?y{84g}ZJ`zrs=gi-b1>#zx ~4.5{&%ԽwTMON`z?AO?/~G|%|Z4M$3 #f" (e$Դ iI]*OE K *i4|V7g!| yx2߼mO_"bMNd< CO$=x wqVDz\xG~8=]3 sMg6r_"/W烣\ċ>%OO%+^M؋(t:~H3/ǣx1? xKۍT?]/^T˭u?ENNuin 9!\e/[n*QW`x^5z (j;]j~q|{~9lM;9B8i..?[qLw9̣}Mò [Vѿw]/fܟKxr>9+6ZE]:Rql-(u!؃|Q.n֕_8ԇ|x@Q0Z½"w;i/|{I}.y\TG^z(ʵ,}2.NB3>>*͸:8Tj6^{ܬ0K8/Y/)8 OOY}O[v|[Y <79jeڃ-\ ݰ3eW']tS@,R&b]5 嗖p1<'l2mI{C8ru%nwʺRV|})2O-s\3 up1* Q G-|4J1yY8˧{_TP>c&G.ʝDP,uTKEΉ4:͘J;¼8L'2{OiO~km%w6ҝ}9}cށr ½Ǫ}Ώ6;w'u ͌e|6ơ ÚB7˴|Ң%*iM*]Yeו'tigZ-IiŒʜ4Rd5ka AK~HA EVx`Te tG%˖.xpR!"ehz)PH,+]L6xeRO>XKl}/4&$ 9#;DTq5elRA2:yiQEhO9P"KI!0O}Hكy)#dph,_XQYQc9PM,<\|Xj~+K{RGxJw +Y=N]{ɍlj=]t5/+EQypМNa7k1hp4HqSVBq!rP])1p0!qx#uHֱ1#P*$XLXV)zƎX1]";,E$2b6ϖ7gsM^]|7~sBp0=rĶHI &JgT(qX* 6Qb%E O1=pg6^ r֘ :!#yNa*צ"ٌn< +s_PXiaԦjwu09 uH]E?:x !ZRHd):jtQFaJwFRTA!Ȣ=&Dx<p!@<ꨡtblÚ/,20 """+Co(ƐJ#.6BG bL"[CIZB`!)$5YMJS`) W6GK֋ˆXL͈$vE>:]qEb7x"'ݲ#I{$}&hM3 H Z%Jat.. (b7<ې$zcFo71 =6a0~H~lwɍKC 5o{O~~< :cͱ~gIsʳ'O0}5oxr~7-w_Q=QIؾ/t7i_i2&%8C%Sd.==bug9pIeM zG/_<*ˋJmrX>|;'*&3I55;z*'75F|rpGAȖ [[rsK7ۚaۛQۛ/WG?fIt .zjjPVyE'YNNM{|c4ꏓ ih6{/*Xg/ξ{X}⧟__=Ͽ/_̼}o_N`á0!zkp鿾iӜVޢinDlo.mvyKЮY[)W~yV7/<&.!*-W&hW(p!]C?ؘ{d~$sTJYDoBE!Ad#S[5k:N:=n }:;*VTk#-bV£rB /5W> O Rڢבذ=t4QK"w8zJeAg:iYNnl,-zN+>NelmMoZ}t>ܐ :aĪ>4QP[YL‿X us)Mxv2 Q Ad^DB )SZ8bN 9ڧxY UeiV )SZa(eҨ<),a)Y,B nKsJ=p[b#Q Ax—F+ AJq)SuBRVj` j|/$y+"lVa޲CqñN$m$=tܻ! 1,4Mʟ`YD mKs΋\8@F e \<dd=ش_}qXXpkBZv}xu#HYzj>QIv '[bM6e<> yj1 qM9<1?Ébed_ W,]7lܼ~>{7/ʦʦn n; o.[U3eivuI*LvxϷmTYy.BrL̅e_8I/WnݾWVA;QO Q(MX" &!R:EJgTRbN+'"ַ.tj03/P-@V`\g ('/NƸ/wǞa>桭}pJx.`hC妟Xv}; )+\-a]f&օ^^=J%WXsoF%T b%x$HZ䢐:L9 jg5[!Ke)<>St- njcP_"q9Y$7^0-d)bɩ)Bt%r@eV1']aTC˨lB*-, 5Ζ o3NYh [oV6\zF罩t-u4<0੢Hw mߜHhP}LM&>•6ֈ'>&gR2Opd:EJ t|{ƅ\$E,\pr=Y2do6e#JI /KXxI2aub7Ix" PK.D*:.ĩjyB iwzׂL%(t4T,,XVNBGTU +[t)9E%&z.}pR+eJgi "DeI(y0$h^^j H\3l?8whBB[_fe L>}W$ե'|/F@R$dcGbGr BieBtL7U:BǮ~8]J]+ЕP]!`͓+kd*th)]+DIuOWgHWJ**xBtYB eD~!ZENWRΐ4 +)++@:]!ʚ=+(O$B*S;5eVѩM}͝YR-Ηr>/[ryB+/b˜0w$$<3׿!G5L7ͨdRZJnZz~c]LM`PԒi+i*4hM=R7CӦԛQw#`N .'Њ lv+վSO$%B*BBWuBs+kO K]!\R+D{*ep'ȞΐVTф 3f++d*th:]!JKz:C2]`Nӑ.KFD\v襫s+ie"!BOi-BʤBWrt(io:GRF ]`e2tpI3h:]!J{:Cj+gd+@k/]!Jj{:C21R  Lvh|tute|uRa =^R~%xPe*JXg9ϲ?^1e6P\i/R^\i<\FEʏx%a\8rBYQ9hfeMbg(3Bb:hu-Q?{Lrc5M$C.yݧbTR9@&:!tWT쪀{%]چSo߹;ѮJhW}=y̲zJjbE(rK 3 3ڙnj~rU ϧ1VyMXLUOXS*e sE酖97-QsE gLd2s?Rݻ?Қ!Q},4,ޞ ۄbB[UO:7^|^sU\m.sȴ9&fu#5+W&erU͝lY\5gS] x4 xj]<[#:Z˖RGWKmoh|Kǩz:ߎ4gFţݸq,-_m-Z7MiOڪt6JZ u1РQ77N9N҃@=iytv: wk_]=n{ˑz>C{ tkT/0V tY*eXM6*k;z쥓6Shk am!$a7LQ?YG< ؄&$ycV3;ΣjCI֟)Bj3~`wcz%4p-71Ԕrzq]#̜9g6chZF9؂$ 27.dE-GȩAla #~ U[ɬ5wvp;8wC{l6n(e.wp;k#F&166뱈5K~?\ܐZ.o|n;^ o|_툽PpǩR7i6t C5l]71k+523VnjfH7ϕFߎmqM.` Rc!ɨ1T@+)(՘3Tc8 K ]!\kS+@{T=]#] ٓnC b+U*th]pBFtut>]`ӑ$e>-@) JQtJ+i:,B+DiXOWgHW2h+E:v+I*th:]JAzʰ;OWH ]!SI Q]+DimOWgHWd+2mB•خjОΑU岦9S\\3J rY3Ҭߌ gV#`N] ;VFeh4={}r%,K&B¤BWVԲ3+&¬w$rSM`M|Se%4[w0^>zI9 s`zo_[!=F|{r$wE 0~A%|ghotrnHttFNos#J7F^^qerhMR jN=vӜ{`: V+5|Ռo]bob0#O0'\bofÚlh2FVRco q#XHBU2x3KU]/^ \PX\%%J*/p2*YYW/$ =dv{îZͬܵ]1 \= \1 S{W`0lo*6v[nd%ôW\* <3L[1iStHAR&0ur+-1#NYbH>驊iV_bHVf{ifӓC.,1ldς%hlfcVfV[%iMa8d [?`3sJʝdn%E=;=`%j_*YٮUR6p!&#J+7pJ:n3+n'DQGpkdo殒\:\%+eW/8Sw?>JMO8#.y>h+ֆ{*|:+UkKw*̲3>ag9 u}~$#vy_,J24*D)H!RsuNN~B=8ѧ2U s_zCo[r}?JRfXu K]&`/e(q\iw0ufooM?>OW}-nmfߦ-:㬣|( mvYNnё$IftF/ 4+56S%P$c=|e\ 235E|8; ;c(.g1 x$ )[X1c2b0ͭ̕v~e\??`7sݣA<{j8m:R1Z`)4?i~ai$_F+=jlY;= I]'±uf/ywlloε N^JV[kd8cS+.Q8ɒ;\#O*(~`T VY{kԴ"w(Y 0[:7UlO6d:eSlzyPbE4YQkf! e6gz|xuFs)a/oz:]v]A7s'QdBQw:,gӏ i誮9-4̯5l!Mv3ޓgqts kutP@e ]։g*2/`MPCg}%Lt&c˜]㻹_2Ѧ-Gص8W6hY7/dzA3x";aD9Y`;yUp7Md]]^p, )i1?dGL!q[KF H'5FEh+s; n`n]pZ\l[2Zg eIV9Sh'=c" 2>baL'h)[Ԥ7%>{٠g!L{)c.~.E ="f"3"2kʴ( UQE]V`-~{~O[WeiՑE@ 38r#!H((3LxSz'XrH# be}0z)#"b1h#2&"]z͜5|(x 8[1 iPg[gOoGgEoV:1q#uZ=QVKk ±YҊ#XH>^؇#Q"!#K9 Q Ƒ6sV;uN `ل=gN1)AU\*V!Mg@h(u_2JezS u=hhWv:k<,E*K䮭r;zamV=q/N0NG J fEl P!ДHIq .}TCJiM'e;Q!(8vP)&8 fj:eusCD_tM&K s0ޙU2޺W,YIfYZ"[gP (* h#w+'AzQ@*IlB)=eZQ4 iRʃH"SKK0kY;R9kNܥ 쐧}4 FG 8"FkAS˸cXA@tL:Wx Mx\ 3_P!Ќ* eKexL*%dCY o"XJ-+Wei풏qO"Gq %i&Eb"'mZ"ЏXl|l|l\\U槟*Z9!XŶɠf:pKӁD8{3ܛ\&:M] 1YkDA1׸Ջ2ezG(-q*qCN +^rÜQf YcI@Cdn-XY{~n^+yу+\7ER Dxm")dt"W+(fyM\4fQq-e\R1GQ92Ҋ 1r-"VD h{/JmVQh ī,҂g߱,&mm;Y*UkFmI1!x4kϷk0>o0.&u٠V0#(p!m|Җ>#Ɇ=m?=(k0K::~90MCiaG=e@,hMk۫lJa^iGjd^ wE%KV*JT= O}30pхǩ0峱kݸ;ua(mh4y'[#Hs|RZ7J.Q=I[q%ۯn;,gCz^' nkn-YV`0-1-AB3-f$V3Je{QP"dVad BbC^h%ل_K)rD)˽l8шT$T⑐(ya'V!̷2@iʙ%I,<'GVCVl ӼK>%0!Ӹ 1/73o]H?OjJ `PʴQ$cZb2$RP-kgC?3`D PMF&P.=Ppbiw,FyCF>S{61nfljf @3s:ӬŠ~v[#tC3PR seHq$a6h+H>7Qgh5?``b3hۘn K \?ˬ˜g=?N@cٽF=.=XSQVFYnGYJaOl#`8 :kN-t+Acn,SA1jJ4fG%se5b)AҐТ9!ؚX6F7OTa9(=Su8}tf<}x_|OÙq$Ȥg`2AcZ*'2 w 1B̂xlDxDH}yҍw&40 (is((J#`z 8"#WNн}m9>f&5YLu8([[bYOt&/طO < hd: &E.zP'> CcEwvM/i 0nJCiֿ#7-H!|{',taMY1.똢ӖG;OXuqϻ/e_Z~N4.`Ƴtah~cX]k(4 %< ΁O?maVx? @8|~5Jt7z R,_Gߴi)jwm~^sjoRG;MÜo,q4wyyitͿԷ/-ꇅ6LֲKk'|XH=)}Cvb_܌~J:N1L0d=e= xRN Eś<L@T]$ϔ7:{[r1 wEXqmX5]3{ӎVJ$EK;*=>kgfm1)i&jZ7籹[Q箷M|6 ,8ы!:mt* H⤆/呋c&6Mʸ bd.R5DY˧DL[c!=ݶlZpDzLidHq嵕8*E gn7auį귒oNRRD5)tLDqP84hɏj@*4~ŝd!pLρ, s&N°V }<)O0k5f,`2b=6MD"2 J}H!K9d~+B1iVy 0*HeC,(<'Yj`HjX @ g'#aE"ag{ Ƃrj6r/P':Ek2ɶA5} Ajv}w5v~qV'OjɵuKuoڱv/m!T {=8JRIT#i,\QO? ƃx@UgB})n@ (rK%bdF/`993Tnɘ1Mf-ebpc܃,Ŀy&Y=2:Kovz7u-Ӄ 'H ʉ2G:H?9awIaI0DJDY"A"qKXH.(W,bnÆv`a!D[I&hU&W!2Y%V!!SŕBb%#1S,b4Ij$fj,^"^j rqw9Xg6.V.rrq81Dn$JeQH+xGA`8FJP$xh5gJz:}Cٸc[y2C<܃[s9Zȵ s4d$!Y$(nieMI&0ڕS*i}IA TM?q)DvJV1%P |()[ҟkv<ՓU:e!#h+4M rQƔ08! 5l Dԁ(WK L4FҎYp-Z^j=Zll8H'R%#WJb}uP뭽.i71$iYj܀ҙ7i5$6Qr ab|BвZ|57")TD2pṵ4N+ , (O+x]5C㞇*4g\&O-XN B gԫ4JDƑRx˅E8 t`ga]#˨MR#Tp*cVWH%CJsNASDp1Zw!mX4/S@a`  S<0',R- #x#{:H8<\CLBL|q/'ìlq:NQ=? 5= %>O:8IgDS7'/x_vL~ EǩO JHKZ4paoFcD?(b 0jPzogb> `GX#2+ttbt.h:xyƺluWE-6`crTIr1A\6rguȲ:?KY ݊Ň' #t[G?B~om{:99]^/^*)<*#.8L!Bvsz"e\ݛQwj.Sh !;oofE11H}?ΗӨE3TߏBp&Տ7IW~;POOޞ˾nH7EfyNo5ӤMZ|Rﺁ9Vսv/k߻J):MqЭ1arR7PLb1>Ť OǍ.Ǎ _FUsp `?8:{s7_/o^y:;AF`(>ؙw"@xOܴk5uM۠ksu;KhWy3;rs[/ J??|z]GS2z#mkMGJ~d'` b4-\*!S.}-B6 s9o/p_ Ћek|$2A 'ZR*hNKGI }ʉsmK:W֫C 8v!X@ A&8~?8#e^q>Hc+taO'˞A!Bf-HfDHGvgw\SfՐ WB0YUKp|9豇|$O-0 h ث iؾJtsyiބIULd0[5Qbd2p{,Zl݇*N5|Z#l$N ֹq7-nG⦪mv9~z)k|пyյUNbuv|S,TLV0#BXYL^jD b7<`5?7/~0~wwJ5ْ7p{rhK#?/T"΋VVR&&05IQKiJ(hT:vF|$y|i2?2h\{KHX']#Al+Y0R`D$| 7AT2A:i0V\9!]p46$ cXR)07A*o!o7aplPS71υsSw×_"s΀kHC B@\Ha%{FR(fN5I1(.# 7m$}')Ț A|%+vy¾go(U`_LQDns%OjzEkSEi)[8Əyas^Ezgqc"`q-# dRKJxЄ>ev }m&~3l6F3KyEZ:p||{lr{MÆ"UAGE<[>ԧQb0ȵ>Lfn~>V۪xk)+׿_6%v7gfx~ jt[J<@&ߑ[8tT[6]j %[Ƣc\8`SXH]bҥVN2:l^zUWz v IUp;oB!6cF@X1b"iZ+I^ ,:i7IK! (EZzL˥böhmSyL=ME` 'r~)D-?`R! 5P̧~A.r$eE8)Ns-稬݄VrO-_N_דY}>IB%' 07Ѹlٽɧ^vJfb z z[Ia?yMܟvy峿L{qZلA pE*D'<~]]6ѭJhWW* v_}ݟwQGTRJm .(LU&zG;q^~a Zs[vYtJawuwo-]&YIe7:"? N4J 2˚f3MD2P#Du[WYC G$޵,u#_ax;.5x(7X8f3с0E*Di|K++EXu$N^xia?)^h۷ХP)MbiﱴuiCo7DWЕ:] aCt%C ] BW@_] Jg<Żm.VJFt%(3)]!]e(m.ǭЕt%(cV:BbH?4?vJ ] گ]i?eҥc@ɰ]P1(pi3OmC+AKAWlqω?ZJkM|ߞ驴L? +޻!%G]Yt~^C^Zdt.woPɘ7h {_Nwuŀb\e㠔ެonM+MG|r.~S`/yi3vw.(bwQ{iED'W9)P/Y Y Ű<>u -; ݲ03p@O"p93}.?Zrx\sQZ56$ 5S[}- l_哯_o*~;\6?1__n/}s/7p7md;C4Ҭ) {cjlaO3d;'}~5fWK%7=hх4M͡TZN{ҌҞ1v6-ueH˝r~-dKD(c B]2c| -Ѱ9NgE;X]XDM>gOr;㙜кKs2^qDnd˾BIyX*Ø3$B#K \fh-d+YjHyZ ?{"ZnCLn^*d#ꌑF#@g ŀ6nLS#%csQ"π@g鉤y|xwZmM4{t`#:WDyKZ8Z)0<'16eϛ̪fuRp2fd93JZm@r3Z-E~zsiM>lDFSlJ[Isi!s`@f3e mFi.f^ ՐR"@"%!Br z#8GB"gi, ֯o'2=d& Kvr,Ǟ5x0f5D /5.U{;ReX3ҫ 3:S"eohb~ lb kX[ a۲]h A@ka&h&~4ˊ#l|\5 (ΠjҕA[k5PѮ!\Det((*qt2km ,S\D?5qܽXe%C 084 C9]dP5qZ _W;dWl1wB!ՠPwjG`(c)(b Dah!$( 2"*kI6L3Vtuk. Ab΂Gª ݄cR 4[˒PI!0Θ D V r 6|6@@ 7P;+E4GRN8/PjϒYl&"=d̟H_(* <_VW"h"uj{UQ@};ZE "y̼:RF>5ALJ"5ض(ʻ5dYBF TDRG4K#FP>6i1b "hʲAl]l{/뗗~Ŋ4dY1<4T(hH;L'kQ2_yONjc9ԂT7z~˺ L0c63ef> :p2HM.%L5BҵLŃZ2C6a1%O Hv9!.V".B ) >@ E&rZ'd^0P>DO:#./ӆWY(^RX"ɚl>A& q'2Cp?,P -wl@T5B]6!H b#җ a~xߟS~7dpL4K V]%VXA8;KKͼATXDebuj W5M pOk ) 1E=`eW\!n[%[1FcGX $A-D/#;PjbPK`ID[ʼf:G ~"hm5~3E^B`mBgU酬iFJkh@2~ȃ!:8vGyPQ`QÍ=`Qy KiAXvXY%3 Rr1Z7N BjϢ;hb"F΂P"NFi*iZGo ^ͨ"`!-3!ZH~y>X<]NPx'hI@Y@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': Nqnk=<=>NLJ/׏7hw_\?i><JnKƥdק6ă7.e6j\:gCA{ldyfJF ] ڜ M JWGHW; _׏$ <NnIs;y&}};R^"f~'eoOޟz) ?nRH^Q4m (oKyC4-7t1&:U)hiZPF4iz+m5<Kt^CڃHCosd%on[+ S:F6DW^lͨ+V8t #Еf+t%hs>ttg=Pxeӛ]0DWkkӕ̪tuޒ`Yj)m6>]e4vutsvKt%_|#^p[+AKJP&-eDW8m :+Aӕ$tutmr~KpΓA%f6S Zd+۝JWGDWwPH96Hiq7Lwl6?9l r/sݹ0dz r&(^on7c讛beFw?Ï]w]^]x1oٵ=5&ſ3ӆ~f׳&Y 7k+ ]ݯ O+҃$Ufrͻ r5QNu3ʏOÙ ;GQ帓bvwbĜs3t8ט$O#+<{_wTrN6*7G7w䌎W%Ip;}{dj;c+<\ #2r"˯8.f쇖eA)'ﱴԡ+Ly3t%pC ] 2tut% ѕQW=Yt%;ӕŭo0-ez{^Ǐ'o~qe1ŗzW43ڃgM̴`D:6Hs_eɛ2$d6! _q}Vs 䊈>E?Wߛ-^ҝ/d{ϗuv %G=,z$T3wWŨO/+ UXblMGG[`@& 60L@"sC")I!;Rr rx <.ׅH ̀M2Ύ&nCR  P0T@Fj*e;vA)ıH1G$xTjpP~Y咽Ʀ"naΦ]TU |I+0wJҠly}x BUu l2ZwWe4]81#cB2tPS Zn5v懳4˳tk]t  R1uFIp)u* 琉Iǵ+*AB48Zu (WKM)kZa4(K)µ ^g5Jmpnו&H) OQ2GcWZhm4 ;'ca /H(\VBfV80.uVK")J E-d4&jci*PP[ Rӱ&$Ry `1$00-Ip "lXĄ9bӱ,Y5ߵ-=$=Y?59BƔ_>Upzoʲ6N w'ޢo~Z-܌o1f^{;-2֛8͂h_L&J&Ї|88\OeoLU 4ϧyro< ?Uo? ™AkC!>;SJ78di/0~"y PD`p/2#1.Lk(Pe:C1f{E&>ܬ{أ̫K61iʃq\{N#`0$2f0؊:2ô X 7ڈK=0U# be}0𜈈0 Xy2Lךm16@1&")XmU*v7=X]{u{GىhC 2EZZceiF{!8:}b"{&CD )1XDX,-,xGI4 WWەxmu_vTl8|S+]_D ժkow}&ߕ NxqJD\_rZ窔PNPi03,,MQ-b %RIfCYQoZ~ %pP~j|>1 ZQ/=ctH<eۓTquL8m232iG](nOf9+j̤tb鴽JӜ,`r)dWy`E=Ge{-7 M+j;|Zjp=9+GIʎo//Ts tJޮWaL>Qv2a-ŷ;[Mv9 4V,^&y:^'.C/ŭI?Tr09ZHV$([Zv#Jn1 އ)\K^>C4.i,2a͛Y+8H)p ޛ"Ofa`)ΌoEҪ WLEAoPwqtN[~.}ڕFSV͜mM>Zk5n'c(z/c7η܌Mgەtl?/o3ٛ\ԗ^|&[v䏌[e5 DYq~K<^MWpJ|!-x*])oYb@򥘨("#sqËVz¢ePJřO ^tљFg^;O=EPF'5|(\43ᤷ!m"U#$sA AXe%X7,+|YA5M0_''&FɁԂ#{-fJ#C+$Q!-"N8)a )k%FzH/I<N}pvmFw go/hf |@y>/5څ&oU&m<7P.FphOf2mF \-jEְ,&(ѢIA2-44:G[2Es޸SEN- `U+F Oy;4B`ɘ%.XL -f}-e#Ii;cƌLZFL&Z܂EDomp(FnR pڋdtn_g"81i<lC&ŒIĂ 5'ת! ЀPxT ' $&޵q$/!d] b?B?e%RKRvW=CReiQ,ͫkzꪲ{A@`ʂZ͂m@3#X۾VmT'^d1Y3*ta5R^YZCn/rЅI΁X\dNɛ/qY.hx5#)3s2&7KfQϴRX+tV `L*`2!EQR!k%ɷc8s̢"gIb֮jm^YkAk:.(r]t`R+.:.wS"cV TՇYic*Bȋ 隄,jF])"9h|٬[~eQhjF54xC̴* P6]$q RNV3u:Ӧt\)]ֆd@gB BiDɗIm95wH:^)YKՋ^^CπI۝M$/K#tY &cc)&A/B/>Ca;}aSTѫb'=vi]˪IpCe?J$"F8;gu3ܤGeNʉF{kBj$$W'=2%PN ܊AJff;I*9W2qY]N/$uTF `T'H H֭D *Je2A.K \KT Y9֪ {,ŵ6ݣDrdAd2)4XO_hbք|ɲx ~c\͠YXWY*-B$ROB&#SL_u:Į -+ݴ) Qd`0@بr֞tuJ $vVh8WyM`ӸR$%yO&Aj u 2+‹䍆BǧZ%4SʬoGpl7l,,hiiyHH.Vye-G_=qU^hb|(\OS ގb)$k;)UGp-9% Cc\2qpNGwntV i;OҊݤ̗þx2e 7 3)`rC%]&_xV4GIيo"JY}rG37,&&5Hu0dXn{xtX%ݟ@hA$KH` M)!"(f銪B[":Y;i:y )dmT+@U#SdH)Oaq ;idLs )Dɮ=7ưp] uB߱>̥H)O$&I"6F[#T P|U_iVؐ ޴Χž1Ⱦpܕx <3(0g8z`nQV+o4ʆvFEykB1 h}4Sh$CW{"npMmJfϊ}/sT D7G'-q`yY[4# E27C4@Miȃ? B9 hcp.ifL"!T 6Ԯ]͠B#N^QKr.r=~Hq,~b,W+ ]F『9:Jc2)&_.ݎ .?5V.%A 2k4aҊtd-2_,kqvMw|qx7kGwWWeXaeew<{=Cm}>.}1:;g~8M8*;hw8-b1?EH؆OFw$Zݷdq)x?݃[w`||$nƾw{E'VWAkYR0ךm]37]c; BZzB\&n~8C!vmZ(7h|e}K^7bL~ƳL_$-ݣ.fyrן~bB#YV.6rx &YOo=R&l XomJA@* 8s78S=Ǚnp3 &Ym6'$%)KVhgN'g$3,]*${鰷#Dt: JޅT Ykäe*9HOm+E> \;p{. }izz>IZMsZW fOW2VUǫ]/uz*C,ydELqr6 TBL@rBkIBE.V:a #92\653g=L}yL< qjg7>G?j3)5)&H͙0:f {!7'F(Q{t dN̹2Ibnl@:3K<1`.#"P,V^ (W)J=pk'5T#g3^!8Y.ö{ 't)8#Uͅsӌg_?՗k*Dѓk(1v %3Ai$9pֲ]c2ZK. [Ex+ٿ|'>_ahP_j? {7`;󲤙Rbv~Z ScM(,5(hbi jsͷ:<߭^dػ<]޸wyso{\lr) A*ROOqIݡ.Xr嚓^TT*#Xyr&bb A;`Z]zbsY5,Fbhm&LG\$9sY& qsȜUF] W~t䳔XyPq=d͍v=PS3*QSlJrz.%jZپ)T^S2fJHj?FO!U4)7$wq*&3A^'0Z.7Zi<&DED!A>R.9:˒Hj&A9WM1Q,,rpR!8\d3r; voj/҇ $ڭIG˧nρAR ޞ?Gq}Cytin?Gnp桵O?!Y9w|=\_?y|Χs>#>lO[1>b=6͏Cs#N y{ _ENhB8mJf*w*ƨ:ZZpe*QhI}*Q(S JtJێt`g|7\fq~e2/QW^9Jp9+вJ(MXu@]uGV] n?ɠ~뻌וP% }_YՑ8nt\ Nh\וPUWKԕg#]~Ar&m'ίZ)IW;3woݵf{Og޹ zW: ɾ?*/n(+T{FL .=3}zg-{np>ע'? ~ ڙW|CoVin{WD8۟/߿Ol)=?h?oO6ڼ&:ld?4b;5==VgTN[݆*ӆ,ֽѵj8{jPvroۭ(^e1vnY71{Uɻ7 |D%d?ӏ|ȶrtK+Fѕn J(iMȚnd0Ӎ\/ZA~R^GWQ1AfՍ7t]J(_uD]Vu?)1Acs w;1n~y3yyN칮r/ys'緹H<=$3d޻&Zn1g a>W*&Ylxfax"tf4U1יVHW~*] JhzueXuD]MAǎt%.v+}(] D0vue5; .w 6GWBit3Xcp芌R] ҮuQȱzIޑ+5aJR+Y*t`"ݍu+] e+6uO72nt%zѕ+ j V! ]JWku/o=yͪ*Zh#]8FWѕІ& ʨV]-RWĻ n;)ζڞ9'ṔmΕ1N1Q8ܩ5G'Z'`%슨ijUmD HW̦] n A^WBihueCUHWѕzۋ6J(^u@]YUP:åЋiʸ&K9Xu+kp&Z}t%a+`ÃqJp=+ nʸ]-QWޡَt`b7\Ӎ6>%)jbiJѕseF>J]-QWS#] ԛuAfu)?Y€A*x3.1phN^4-zIP-S(Ըq[ :jtF=>O֍>٨gT9g;֨ݺT;ҍݧJh?QZ~RŞ+$4m1qLbVGgȥb5~ik6&+/܌Íԋ@K]WBIj2ȷ;!nt%ĽJhY]WBV]-QW#]w'+}ҕ'BIvuEA{EJc?py'PZ=w] B_u*riJFWuS1v5yGQzW]-PWxQБ6] n?sWBq‹#9'\M2(a (YsWKU3Xpwhѕz2]-QWQ=`u7m]וPujᳪ? Y+J嘳c?Օ1LtpjާQȿ.?W-ؠQҲ?(y];=C_uIw)_O'Qx|AM]ߠ<FǣnOkHIxJSmϷ9Xcg_K?~:}ٽ{C{<՞AO}ٜ_?ʨ*lUÚ*'ʧ eO?gÁ^j3q1Gg*# *-Ntב==,χ,;]4tu18ovEjhrʤCESV0g)՚i`U}*/ ԝ [;cqN˚f/AU4e9ce ]2iPYs֎Z-8D(bRkh&UԒ͵2&j4jAۡjy3]K 6FKcr~;T08Wj6֊5 kg3")טkRSj}MˢhK!TЙQAʎ֌spG C͈HSZC))ϕv3-Z1S$3ju)h6f`<,O+KƢ5֡L2:U(CN!PUΦ: NO!B #e"#)u"PoU*T)L5C1gX2|șbo>3ĩy^-hV-MM%jJJ䆨s2JJ})}XvJ7Eߢ_NR -BBǰ`O(kcmB E}3)aVЀ%B![뚯͸\A@^X'=Xl)6.(C2jP)NC,XBM˜3{"S NZZ`1Z{j)atY# HeopAh>G)O K1(xiTP >` i@av޶Gpl|J؀HV;;m 4b!DX = o xPpxiO; J`#eFHWCHB+Uʈ0H8" ǘa#"h҃VsB( `I&ƴJyMA/CB6tq?,1%$ sb_6 !kڀ)x !#A /=O"j1Ր`ZD[EsaxͶ!VDT&"0~.~o _q4\tRMF)Vĵ qwm[vC%?ʒ$+\Rr,[&mS̝3y^a+= ԥte:H /v @nJ:ȻK\|0%8B5nl*Aᡔ,a+:DFk +]Bv yP"#mb !ݩf%&TAJT"Н)-@Gƪ@ Zcn3X6 SPZu(!?A!"#wGzS #2U9\J;aW@X!dʈF` A2֦uW5C-`J]Cs&4pVԀ̬.Ix"M#-z",f#XHY 0 $0lȾϛi%&C6hijh[f++dlw`q9۴dV[PfRZ.:j0uh6V鲐`3 ^ ]663%ݶ%]f먵MhֲVtA.SVS:h4v,AI(Qo @r2#4 QzZAA3` Eɡm,C1#KF"vz\]&kTP=`e AU )@)7EַYaKl\!>v ZO"LL !6 ON .9-.zW##U@ #R1@jS%cGQ2jKysRu`h\t$i#*UG~ "y]!ǀ&m]u[ YHkTAHRĦT$'K &Cha+j]WE t-[hy~kPAђ?jI7labHM]i*C-2mP Zx3XANXP0h–f|єHBZAS1V`1G rBS{p2qKz MѓFn| pib1[Ʋ5jL\AC,1;Vf5$^$d$8D )uH5F:oe%B0twE#Bޙ F\B[rGAgfr,n5̻Gg$X%Er$>J啖Eޯ+$g^?IFU r_*G,@g߿>Zl>Ȟ;I=`Mb 2ihDOt䷬ru!4Wǫ>@PzS12lND*w,[D =V8?bF%G.t\m" ng×P.Wi]!uGB7(zP@i]H.$Ob'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v;il@ީ{ ?B`AN > q;:0} #b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N vhϝ@fK'w @q@ۡ8Ci~9N N{;dDf';ӵ` ۨN8s}a'; N v@b'; N v@b'; N v@b'; N v@b'; N v@:'Ч;&;z5GMi~z)Y}VO.@F'?%u7%nKJK!ظtƥ^B#`$7>KBľP*i7'"E"mPNW@f:@Z0wGW, ]\BW}+Biҕ>mC$6WW+}RWl naBW6}+ktu8tU蠬]`uEpCo֮V˽O buutEϧՊk%zCWWNWrmb:W ݟd-+BxC`|! `+{CW}+B;JW~m9~]) -dj@Lᠫp2( ;z t_٫?ɦywjDk9ߑ,2Uaű2M_ozb>mgrɻvO2xmg%:l2ũOM8g#'ra=c 8zH s_KqXOfs_U>Mo9 x0x׫@+]cBR zj\I'9~Zx@#drY?頿Ow^=QV:DgrSY¹)M.x}=6gh:6-L>Z1az0ݒRVK[Ј\k] E䴍wZԳF8/۳5-Ege]Uyʍ1Κ^6!Z&MFi塼ʘ66ooۇDZr"zoM2kirSW fzU?e{B6L//Gi+J?S_+]hߥb>7_w28K&h;D7eG7^A7Ⱥ7\YS'wGGɔ~T.xk{"|B8jvy4Q<,,G٤TϳO*rtrt7vI_%)"@Ђ*Z:uiy?8jeGBږwky0u~Jӳᜂ(lso[ῶ$mx^fgOX&.h?*:˵}{qH;~Ηhhj- A7|}m( Տlh55^,-V-/޷ֹZ}+BtE( tutld#"W Z'}+B$ҕA>+V}+bTLWHWgzCWb[Z;]JklQHzDWl| E_ h{jQzG}֮X,\כdІt 9z{ Q+U +Be൫C+}Z`CWט-ԾtvutL0kZ'CeP&x7Zs&?-B4iOC8.=3s+xlU\ȷCIUn^-`ocڇn8J"J;]Jm7>Gvp ])oNWR9+܃ɷ  BW@iPJN FZ2mwޱr:ć-#>=:,'8xr>7+̥WIqQF*y#T!E=e1J>zPֶ}ú]i;}?E]xh#@?.r3ejJt#h/6L UY52/Q)*QX`|ο% 5q|, ۴ z.y#n-eG/[!Z弤Bv j}Yʷ;!I2PM.Cp wA('H8k{_u<>%m46d[Nrv UI-鲠G~X=h\ cq{J.?OS4*h~iϢ]]$]rz5̺|gcXJq)m>­׋bΛ7VJ4|  2;0g:q/i+|n jiiuEN^,}8?.-,F eM|Yұ M-RMqTon}-,5y]0 NN*k5lz2Y44 B3d6G[}CG/NKh tJ\{%}^jaҊ"ԡz|(| /iZMfMhn:>E&l=^ɴ7vvK)d1\?4]me7M 65 ;vm7]g3:𼜟9^Og|>(W9zCam̖,zs~~Wpsy<0pz2oᛋf;666;l0'ty9M]On.J#zC_7n.08;m;ќg ]vxXޚm83Ǚ"Z^mI'J^;xN&tsqtb`x. LDOQkHtGlx1pVVҝ׺`\Vy_O7sϹ=zhuN>|z9SYKA-ülUC~\wi7|92cYcc)WZ}|gQbNsLo 'm{vSVO^NN\yVu S٤F{OS,7KHhSu:8^nIOk0؇VE@Q)"s? :gPRcrtc;(P(OQAl4Hasn dCcCjg|'Ugh5[AݔZwC\\ vZZ lFSȾ4l8w+]-b)u^HjlMjyKikiZ"uD_diX6/>fm^G◴UQ*cE] 44WI5JhK%;]+UIT'CLn_hB,p) e,J R41%5VEo|~#|X6Ycv^|1?' ՠ7pNV XQ1Ž3bm N΍}a+̭4 14s"y@B :v?\AL^fҭ^Mhd6X\eIxB}ȫ/[] N5yl3Ae8CpqV> reO֐eo`BFG'SɪRbv!K=j2˫$ߵ[#mdJ1"@Y&FQ%2I T 8aEB0l {cH(FJBIT*. (R \x#@L6s-@еaؘ8Ɓh\n?i e1iBd!GsV<0ZoyYIR;ʹ/}k*&lFx" rsBV M$ HGkSAqMĹ[kߵ5핹h Z=^|M* cbZube9KU7--w}&j qxI.wم(1i/*>Vir6Jr@ybvݝ|,ۘݛ1)ٚYK%30*cڀP9u%H8d:V68ĥ5m5&\5&{E[@`yQj0Ft&2X` 1NBG4^LW×<5YX%Fe*Z+P;[2@K=wh >ǹN\XgnjcnVM ٶH뙐ڻƑV*/KD9J)Tʞ*8+pb6.-z\ѹjJboc,SNrpNdFPt(2Q%B1RLZnb-H!% KT,[|p,+gc3IFT6 9yAp#uΘ )"w0h0"P#X@H̩ȲeY+%  Ûu}!l;kn=Z啞OǓ*n[m;o3F":[?ӑ;dC6IH&b')3MGBI;ѵ3OORkz֚|i`}63Fc##*-dyN*Di!6Ľ*ϥuBj}cZ97wFke - P]B 5zn!Ԅ/ <-D5wyq!J[ ,8#BHM$/q^D%}d57[UGj=]SaT ~wᅮGqx8a^9GiO^ucO{%b}ߥD&%dAٯ?J#{/SۧO{}?r/{qs{/?b~xUD/ ߇W -Pr.}I s}Z~z3Sհߋ{gOOyu]{s7*Sk7DZ~*u(E9Gv6J)L IRTט5DR)O3Lh `&/OC2R  4U߾u/u~xWu>:w^ʈM/5wkV~[}W݆5Ԟ'3jK9 /tuhx:{a\ʻWKPhxFfU@Bk!YRP[IvekaKlv}C1*ড_YƇhw=501L-Bf<`uˢ[o \Rx*dJuf8>*ofպڐ+ɧWGD@YyE]j)ldo[<,yD b&-*MZw.o[3/1RR; ,O?F_bܡ뫈+,ޤL/>$Ues-Ӕc[&R/f+(|1ڡp2ki+]AU-hnφuv%TZk*2~1- ӒMԴ"sC^}VJCOD{l;S\r{zṏ]wD H⤆呋c&R6*`\j eie|cVAYt4W< GԂ#{-fJ#C+$Q!-"N8)a )k%ObNā.0;JίxcÜ__ Т\Xj Eȅvk?~"oh-{ MD.-q G];On|yaņJ @$ׂ<'LD)Q.E)/ɵt2LIt|ל؝`kҾ8 WDVwi@5%.XL -f}Gh5|2LwZ "ⵌFMFS)5r"wiGoF/G2_>E^bǤQZ?+2fjK ^Ȍ ^t4sf"T9둱=Y5,l36ʫ5,KU-bfI2^]>,l j+N/(|wr3> ΦgT& ʤH;JðD"L4H<$`Fg 'H ʉ4 #ȁRi;^zkBmn0 """iI"x  B@EmRD#3,1hk  @+LFHȔ"M>zHL8cu<XҠ蒆0I fc-#bkGO.Nko]lK6E2.;\\! p&Q*tE ?A; )B ^Jל)i]ָcS| ;|}~"DgPvir2':Vdf$C/*dgNR˃j׮ d+b[@b"bK(F[ _!% 3Bsm -r!g"}a\+,j/ǫ} 7 ȘUB 4.k^^- 0 t (T r'ǦβT'S0тتGzd9\ə"c ԪKU+ܺ'[)zpѻRv#Kf`}<7OT!4 atZ(j4 mf #,dc|쵲2aD$jFM,+'Rꄷ㻑dy9h:Zehh@۩0OE*̝rp?l@-`CsKFDUw  ͽ76RL6Уݲ7IoI"0Mʥhd0 H1z Mjש ~8Iv绔Knj^<:p#.&z]×ȧ_2\/qP iAh( 4d+R(fN=I1(*#u[Ioa\Љ6|_5[).(8wȷ~D B7̜ Ak7.eʧEsPncx8cgmIJ@Cϵ,wJÃ&Ny#u: &MvS#NZh>)lJa^iGjd <!>x+vb6G*r(7ͷ:gV# YgN0A:~5xeoC'8޴X6:|?}.q^Hߕ=l> )--Mg 0)j(XFKi'R=H%(2蘎?VE+nq⤍1+Ŝ9s˭'?5be*?j)ctjA2A0ɵ r B($@bh՞s:FY^1Tk%iH9k)=e{@wuߺ("-!3=Wu[Sʻ ߉zO$i5ݣ)@do$rڗ)@Ddu_FHGpSB \%ju*Y#WW_\q7Dn][r^lOaC,$ɲ FXVD;#_hVЛ >cc&(y|͖k|v6(ډ nϦfbE83LJmס ˧l4 ->]'oTpq5qk[n}߂WZ OR+CS^rC^6}6ïjWVV3Ų_~^ bSEsL<{F>{G={C?f8b6hzsQ(o{Yp#K`RGSJmΌ6> Dc~潘:vac8t7 h&]4_'kzb=?d4tس>hCcn,SA1 K8VS18(iaF 9%(S Z0'[,kfk1"G(]ћycpfW;Ldzc`z oryqtcJ8q,YwI,*XC%(ePȀNI'<1 3c!V!h/@(DX;M>(J#`@<1̎tYAp,:VqFxgS/<Ӑ~$NO$U/aX]Mֻ"xW2{ <x I iŹibE Ͻ7Ђ9x4MW׀'ȍǽ |qmaTܸBYq/K={9߅}$pв~lkUp]=kTV| }tqϟ^]γGǩ@tz6gףa<2{30CWTfx z;1sx48sO fh5v69x}KjK6gEh> .}=70^~,}tM/5wkV~[}W!h:̣a΂ů7 ]]>?~^Uù(}N|cÆ Z daV]i(5a߮vҵ)ߵx.ôfQ?d6_[xLUz(Qיh?Oj]YmT4Ӫ+ts".]-3) :c;ESXS^d@OBcabaf՗Rm ;֌/|}Ԏiϣ{whYky޾f텏(s謁>?ٱ-u%3=2 z4]AQn6J k\, 0}n*eRjx{6dEUֳ|MePoV5dZU$y`.W|V׫U1& A-j#tB M=G)ܸŁFg"ń13&AI}CcL=IW/N])LK_ԃ_S/"{U(2;ʞCh%LAd^w6ӎmivp.=}l4Xsi}nIpCSڰdm+ͳQ(Ow61$]zչSgLBNɇ]*֪|! #3&$J%̍iz߇,W Y_Uwg_+f~B}tc< _۟-{7ogǷT61D?\N_  |~?'gͳ/?_U'>opF5#kG2 s[Y~ɵNO"u}^>͘秳n֏ӣ.orݨb5FoT/}5ͣ2nX<0?-Gu_󋝎GO?X_zÏ ߿8'0ˉ'u"jUx<oth-k54Z`hi[o0-oy͸xd\\ۓӴr{_Of}^/+ufdF|ZWuc',v-=q:%)_H圇Tz]]z|/_Qn@amE/Hrm;He %'z!Sު*9RTڥS3՜L8p6,]^[ϗC 8vUP)EAa1I$Hh V}e$K)F6j㰦dM*B6Ӊ-T㊐ (x:M3<by&AŸ;~>o?&jjEM'\rN{?9YzﰐT$!8jrB'E`I("ὔL)m2sFc2 -&-1RExm+ F:K$5C(70w~C.'O/'B Rd [%29]Z7l%~ AUoϽ2u?M"}B;_cEv c.q-`07!K3=U u*yhP8V/H,#V0Pb$ƳQuսzWiλp_-u绦4ml~e9;[inD8< W>ڻ^g#^դH =&1jjMPrN37UQ$!0g$W'GѧB$߸:3=]V&ew3K9~oES2v5יt B@h*;DI euY)SN.d * oK&ʭ-(\I */=heklPwdɭɔ*ȐZ7y>qrq`YVz|]I=dY9 xR_*J):V;(yfO=TY˶&n6_; 8$*8ʝ7\ LŘ8P=t:FY^1Tk%igS{ |t3gI{}տn\~q zQ +|zOu8W?SNy Nji1ңam ͣ*""xBd BKMgx+ʠx?$N2BAFvi Dj'*" "VϏe&#QHYu2LwZ D佖8hj4BZ6kkt]&+BMɊTeJސ |t`xvy?-#XDb)Aej'pr咕8o4;H5`]䳼P$n5IK_; ؔ۴$M܅ޅYyt0+L,=<[OIR pC.5i}oe!6N)%u14Zu]L^e3.uuQ^= g|K6˃nz]|4e-zE;iVy?57|KW`V#PGoΥ^B>ǟ5]vRa9$$ 6^Ԗip=lFr 3WmVڢ]8; +Fs''G tA[aD*>v3G5*kBkM:PUAsg 11,%! $5T!vh[ܮ7gǩKlJZ#7U͑f"{OY ,xW#G_`*wYbH ( E )f?zHf+0Z/%ͺm5:Xb1Քh,$JjĐS2E'2kbۉ'orQz'3u *dn}Pry_3%Fā"YUR< G#,h6[X4iAN̆`R{ R̡(`cЎtYp%1#VBlnj镞ɻQRw 4|v@Kz7K*&_m׏lx3Rq$OiFR#CbEx\#JȨȁA+b ;c$.ؘ-O[Z0zi]?# [gxI"h `.,0JXs@!c*%Dtִ~#nҗzZ4 mB uXV%ԆO \7ܨ'ĉWY+\8q,:&g3~w3~=Y<1m>TNoJ:aQf'fq;y]rw҅Z0،x9ߓHO/_W]v>b( ҏSlzc;fb@ߗEpgQ@#ƝhHa K  A_BRD g DH`}8h2xX_K/3X>{#w>QQkhpYol/ x25q-M#:}_xwkkgð7k3 =82wb?IUСKWRbG`_9T> JMLm`rmU2w?naz4}nqg&•zA 7[||9.?ީ/ÚNSG,:_KC]?J}TW|ڟ.Ao% 1lZHV~.pK^Yd/pb_*bi?e!1liE0|h=Ma?exOVq; d2iv:uw+Ή:oԥZbMz NF(7z䚗qQ^\XPb#60LgBqף?5E>n 71`yi0$Gš!+ч^m>̊\?b 1^LPv/poRy] 5?l,MJ ,RM;#*hWwmj[q5Mj>يtó.y峋&D;SP! U>,JUB͐ZQf\2ZdaI=̇TTbl| [86F"(+ؚ|lˈeF#L &P d1"Xsq5T VՌ%oiCWhqW2x&-P!ndmI/Ho8଍2#qF!PK|Iɶ([E9I<&GiwX1"#ZG!1^\[g9KN5I^ &!̇99;ڬz{ YŬdApC4zkУ$D55!|IzgH C ] Զl'Ska ~bQ3( I2LRGQQX9e, Jl7A*QQ*+1HB$iX EhZ :eU.$xIkg(%v("QOޭQqI7?mr: Z gMI:#fXn=_ᢳ|&fP"KΫ"ŋq`\gbudX`OBcsbl|E:ΒYŗ斖^:Eat(Y"+ c,:PYù=Nx;(ygg\uW)@!-hHCb )V,(5&A"kG8HY6H|d9`*E{]T`)%g,ⴳB+Ź2Ψ\DƱ]J"qq䢤% 5HF,ʃ"y~ǢPjr>0f(+$Y$ T͗MzB9Vȫw1?W㵮{ N15sRo8s54Oo %vQMs*\!"娭85{;SP EC2<%G=.Ox)]M/?3Wϯ-o5ZGRU Gr@Ld{rn6r$(/ԭ\[{S }nSuuihy _~8^)J1<݅M.u˄Xvmq(ިQ7]uOZÏٕ7ͅWW_^U̾q72w+ͻ]qTOWDQ/,_ckZm -}Z׌Xߌ^̪͔w|,.ŋ޴94|sf}=ꢓZ]W#dVwϋ9B +:Dz4W"8G:9wB_?TL>{h<\ǿ 7?o_ޜ)>}87p#. G X3_wjڴiɘniloѴAӳ4mz7hW6i.mQ\.]՗Wh)bzxFͪ.q'sϲAU?N9*YRC!V GKA%<5o)kBEam`nkO~gn#Yǣg)$1͍1pJHƑH?#u̸4'I)[;W7]t8t@(VX^0#ݏcsSB  ɽ╕"6 {:t*#dk2eh9#gvg;<ssZ5&l2LuZ ɖ)Ø4Y+52*Kx:k+|{ ;<[ayi=z-!LOdkբĎGA;:v8SloQc|(Kin$>__m\JT+RӨ#RSP_R07hFjF7˚"j۫MwO,Uiyt7 \NWz^]r!W7pU>K LJ\*l!KVϸ"ܸO%7,Y7}q|yCxݣ58R05^XفJggsJ-\KZqjmkep}a4-$5Qq36ן/ڨcM&fEtm5hC/ z~bvձc;Qҳ-6[H #)5+ -N+A2bHv/\J,y8~bv JޅT YkIvbԶlrUnY3iy<9:rt?Һ@rG[;.c hv9+{x3S^WA_=)b_Uc'?j-)eR!(SS ⚧ Y9P[ ࡒ܆R3 h~݇Jr8*jA&A€*6LI f&Q0d8؛W4W"+չ,={;8/;pI5kM!Ht9$'aq"`NZU>+NEoSr|?\;.A]&Rh@&n\|yJ/tckD# suv/f閗/~o۫b_c> r>$7|nh/څJmo|ADdJrD"nY܂^_8/+EȃWBv z[o_lkn㵛 o\nMNg_5~٬'+45k ,jq\^M`7UFf9Ww-+U(d ?FrNjm=VX'L+j຾5tE(Wwz6t$]`'zCWW1"\v]=GRܸbnjIU]!sQ;N+oc.3.L@T->aw{]59ݚmQ5d'`GRH2Wn~}v~#&%O,iipAVH|J1Oi;Cϥ2\ծ/tEh-tS]=CG^⽡+e_ u"+ @Wχbx`-zCWװvE(` >Gk"7+B Q*]=CR7ݱە]\kBW:]J>ϑ48}21 }+B:"F t (!d;BfgJu" t *Ť]`ӟ+kM_ ry䃫9%X!*!hK[BPhdb >;B)@ϑRF5̚a~eǔF_ٹG|w޿SgVBՙy9B4/-7 ,/Ԕ}U*G ٕᫎ6DtJmɇKƱ?4 .8,q#]58On1O1ҹki ?~r + ~Dv;Ȧ+܋lʥd!EdQdFր:Z+NW; 0 r±eU2Xy9"suP+"`i SyQ(rn.p w4s'V~\q? '=vsTl]!o1ۢ++M_*u"+z>t%zDW؊•vE(aЮ#]Iù]!`7tEp ]Zy"ztFZ'v?+zCWP9yrEfz>t zDWL]+Bc9ҕ6N>5M]5p{3HhMAB9ϒ\?v{c ZaNWkg&tetve-0BWvaLt )aV;kRO<޵#" 0-͢6df`I &0u#KI3blKn[@[GcUư1%q)!E 63\BQ*Bӂoi@{nKf'l/]ml-Qzt5{]ʘ< ʀ5 ]e ]!GߍtQRh3jq@t"ߴ*p4AtYwB!]q̀}Zp:htԖ(Mc]=GPqH 2\.2ZNW%І!])*f ]e{NmC ]=CD>\ ]e{^|%ZuXCWϐhl! f ]eo3&YҕTKz8 wlV>ԞQ&v,Jr hWBW6B6fon-A9dv|mys19FϊJb%vGlɡ#iQNªz iR<.ǔұ8W:ZUF8vS1(I n Q@=P;dÃ昲C<%hïG G=yqdGſbַ-Eǽa=OWp{mB^?o?=HSy;fqgE 6]RtC ?FUHWBWZw(부'+q@t7CW<h;]e`ztׄ]e ]e`ֈJzt%*2` CWaBWPVw(lҕ <CoЛ$%,ɋιwb'+ Ū-|Gť`i;*FH*hi J`9 yMO]ߏkcuGhD>D!7?~םk<} \oBO:BAGQ\?hlk;x'1FSI*s! *CNG` )A"5P v QL]U"'QkF /v݁3C4м);y9FdB}^(e}2Y{Xz^0rLb;;{pueB-v6\/f2S=Ҳ&ɢG}Grj~-B%` k+0!Bx3.jYɫ2 ˙y9OOfm;{;mg #lDewj Nc2ޣ}֟ڧ粿kdbW Ny%$ڵ[9$EAZFMI悗>>20!%)E^"1R"Џs&*EɉeRu=I&XA'kȌ ̌@I 1I=v# d`Zg8Y({ lb4./d;E٘_hgYRT.?;YU_l0@Ceua[Jq:>쟝_rv/CoYՙ1#yliAeM)E״o*b [ˢ? MWR)tDHy17 '6H^L(!FXByԚ)c$pqw`p`q. IwA#`HJep֕ &H)5Rȓ] lD:`TJ%!P8!zО;4M|MBBM"u! ©J Σ[[s1 %b M` T[g T/;U)ENntN:&KI@a$:& ?k&@$SLf4Hr$Rc5ұq8gi]yP'3԰6UF3MWvFf^am-^Nׯ?®e{vZ8Y5l<ۚsNZe#ۋ2e,`ڛ^Ps٫(S.eQD :NG2&v lI.hڟJKk컭/l<[6_tsa ˖2ByS3&*z2ZFCwJ׶*fE*WA:S NEXo, )zHQS|e6I-(|!QNyMz'(tܡ&T$UdK )* ʠ6QF+QIq[TM.{:_`8gYtd8XOGZ1\A}^`qړӥX;Z*k1(Ko7Z׏¢#\?'ZAK+Q:V9 /$7x}q΍=ܓ[{o&tRу<Rpr,h P!NȨT'~*HG\J8yIR:4Ă$BzF'bLhmNs$ؘ>{2}4xbԷhkaFP?`R!?2ª( vNBa$e9z^: cK69ό)h((W,)ڊ&1FȅGjlB{vmLDS{z[wgAlo6uy7bMRwwjA$X &cCvwNId A H"MnMZ$|BD y;XP>4g3%-CqT;*E djG28w+S-vweڱfSF=^{TZ"gwѻ>L[쳔wݓ Rr8K'Q.VL.Ve=OAV(k|5 .@h6Q_X^mnIJImTK9Q,{ThIe QP_&D$C_K6mr>ΆaT=Ig%.N:R^hlL\Z(^2I%PpA )E-~!cV5#Vi-yB뤁 +!DmNXV\f B V<mqVl?Am,hYY#Bʻ!JB|WpRaqv#o/_/LhHb$ :2%eQ(YKG'-cjZE%b\= QqIr 8ɄWN TUUpzJ@[͉UN G8nc%6oI6`DHZMC^oa!$$煄 '` TArhk! Ar[Bh )h0p h8H ^@U VAz* BARio;Fe^k+0*DH)V8夬`8hM<4((?hoXڣarɤ0`>(Ҽ; 5i& 1Iۛ'Űh*[ ǭ|re-&0hB+__Hy*qxIZ1pl@NF)k6yxhG<O#ƖUۉJ^wm#I_!㞘! 8yf/伓n`OY _5Iɒ-l9"&Y_UWW ђxk$1NPK4S8-Ǡ!oL*10H$:`,領1?vY40{T^ w[,RlfRM8eO/ZwECv?`LE?$E|x['ǖ=;"H7vi vaR~X@U^p$DceFaF˪ ӦVb\\OΊi-ޭHޯWgϝX˒wKNp?qT'~J;8OJo&L`\T DA_j\g_|QuG|'l2`*kFhSt/[.u%ϫ33kLy64hw]u}Ee;)w+'yYCͮ23|>ciz.M7;?Ogzcq5G)KC.Mτ8 W?_^dЖNdZbJ$RP-kV[nՖZ1kj@$,J`d)'S`0Hkc1Ֆ=~r3l[BfSZ3q4,-2Wt.AOp&2z#)Ν1mgC CDNc!6ؤ^*h26&\ KII <}Y8 nOړ_kǼH0j#G6`:zע4ZVyF?NTwo}Osc i8VS18(+CN ʔ DkY7Co&C";2޹ݜ s q8۬A-cn5?Dÿ>NgJNLzf V*CtLKD$0u0R<-"R_nG-0(`L` R{4A90PN=8"#=~ }mYt;m)MMq Lޝ.!wASG # G)G(^cޢHp}r/G/}4( ҟ .SEny?Dɑf%Ε tR)18b#š@\J=%+td07/ǯ1뜌oO:(*Eu *sNYTCPߝ*h<%U[,6TQqIggeWН_{u>D?'W׃~Daz&-OC󪪚z;2iɠw&g)`B՜Lm'6:xlhǕ\UNn.{Wj`׵-]IW'O:WoZr pR-ߚWw0'@irhwחwt*{]_R:;.AUQ3dUfUAo<L7n]EGL{Niά\ SP|օ;:x_Rm*|iJTu?|Ue~Fն3GjZJ$AptNTGԥLdS[M$}%2*2-gnny~R:0"5e)eJbƀkg{Q]\7HԨrr0ydZvʰta'Qr^cȐ@)k+ qTHxJBZS&Vշ$ߍ8g{>E3  4X G2!aCOĂ .kb1qR$:,{ yAZX,`DGEt0Bvdڙro,(;3b.qPʗ#aSL}xu8Zj~DKOFW:(io +I,|I=u5o?YQs Cp9++rJ8bXJ84Ny}SSuVdJFHqbO,8>jK ^Ȍ ^p4sf"Tnd&vdUaa#6{a{[,'b2MxX&.?v|5vM|Օ_(M7읏'9bkP,(#Bc+ T2R0 !GKfp7 "$exTH'iG2n.Dʝ](8~A<;vEmq-cB[Ta +K#h&3p1!Fˀ'1X:;u6<\H fhю XrH< c}9P4Ogg;VJ" b+"̈HZDlqS0 TPX& aa!DA[$S@&W!2YQ+)=@|XG!1pzKYA9% aZ2#b6q#Y!:]qfEbzGR(_lTϙKAs#P4TٯfR-Q$>j@GC*-eqc4ʽփ"ˁ"% 3BsMИ@6|+\N.3ζ=p>Fy 0{þЦ ^fL`*DbZ8|Uľvfd `"mՈ#G=*Y,eU/XZB[ElxF.KtxRq#Ln|syeO!Ɨ"]ה),_Srm7s-'^Ho5~r ;0%Qޅke-e` 0I򃋚X&HSE57V"FaOjgяW#ރ3m]m,m< 2 Q<[~}+Ge,ZZJL)0"UhV(4dI1EQIF cE`˕!S NƆyKJ=MʻgǍ&8JnE/@,Ccw]|/s΀jHC B@\Ha${FR(fN9I1(*#g6{07_Skv.yh?4_ѬdK ~7ڛ~iٹvRʓr\L)91X9 {1;kH)EԒ)E 0[=>z̝ u_ }_ywS=V2|SXRC!fH 9HysQ 9!%|9M7[Y(5{džN˳/&P.͎>Tϊbn ޛ>뚾n?$:7}|w1nsיyZpj ZFƩTD(wNKī$a~n0BA3_"/R6ffé`tu?5ۻ0<>XpQߓW?OZ+ OJrø$d-0E9jUŒ8sVKE65:4NTN9eY'RVЦ2[_bu%ϫ3 G ޵6ۿ<iVr7qqޛ\S$ e)*iIEZЮ% zgg[~W,縬Sy[Lo0}PӳY-׌'~Qon&V &r_,eP\ lNLj9+j$u .q߀8})pV <ޮ}y{?KL6n0÷yQJy3x6bcG9GGDcH!ZynrGMr[҆)K\QoSv:"!x4;cVql|HZFNQ,c3&^`D:LÝAF^@(8$sc#Z!OCEEwr><vaI +E2)FM'2: =hnt`2<>jDCP3Un(HȲk>|cl>1cE8ns&1U`d}P.g%n|ه;C,u|N̸Z!kQ &Jd4;Hj zM /+}l<= z.i`M WZ5]桽Ne) ?bx4ɭYo'?Ƕ!OHCt^^]=tgXS̚vuoBɴsjWu ~˴As%W`mdoMĂz֕N"XyZyМ6Ztus;.Zx׌8=g@yr]i{6EElb@L HW]i!y1 *dl;\%qJ pW qs&LQC9/}q';&KKj!XFXV2A4#{h0zM05HhVK%'/M6l7G`M$0aanr0}7iYw"]0MӤ>z@;WI`-v@\ٮUm=\%)L\$;W 0~5+ѮUWW %!gWp+w@`U]$-WIJs+z tw֮ \%iv) -\=CSz=@{IUǴgC3*Ұ/R: op\ĀňSF2†gщ,A1Ol9ZS{C !tF;h݇۽x@c ;Xr%q8Mp}u ^\N4L{ T|r/^:"UQ%V(#1Ę1rS)L *տYm>2&__G/_6=SS9܃65L`RGS 6>DcIxe쪝'i]-h0)̒a~9~F;C:?ٶ'OO`i .JKlD2k[ro2kR17rUp>V)Lqח xF;˼9EL[yn1LYvRݳꞲ4J%)G" /HXf4D!(e 8XTTsL(i­RȰ, Ƹ#a VR.k%#RHDc[5&-ަ7+K+Z7IUfRnk+q77( "j=WNXOfhxT!p}jӘG#0(A lcwc0XJB@Hj@Ca nL| qbRK 1RRA9);-,y5>x\#JȨ e1R;iO5̻`c< O2׍qE>e0r89GKbE+Npsa!nNEx}IQDYy,K1p8{ {JeAxsQ2um=7{ ~Dי{}^ŌyEw.84臽Lߕ-|CYfޏL2x;uEg'0dkNl`r]+ۿnz4}nWuoޞz|TGqwC}0 L$yԏ =: ]]?N\i<Ky*)x/s1lk-$+7eWQɣ`=bg;ه:2>D3^vOsfbxxBf<`5@٭sW>{]V}"Ku?|ӇOV8NJ줮(>Q>- ) P$ݎdX^d4Ct91s"ŘZ>uf46 +~\T}á;v 褏a O8X.<^N#`i6 fDOYVR{th5.;nr>|1; κAkcXȎb w/XşddM,#qRqX~y_vnuFsڔ>N&]|] 40.r&( cdJq! D?:٫(;NS钳?ð/y) 8L;Kb'8pxW8,? tU2Ղԍ>1);Buew1 pJHۻ]ɣ[npAoOnU__[_[a_$xM}pon*\=dpФ^csj=;lɪwN]Yqb|JUqݩ.R]& ,Sj)>ydUիsŚ:<^K`eh."ln?ъK3 5A5&h/aj3+n4]oXVq nEvQuM=>X=liFF6ph Dy ' Ah2.ƥ JFޑ[ J`І""zi>gG[̔FJW^[Ix"3Vh it8/x2| 5bXXE.H=={l?/Apji1jZ/Ćq%x" J`+Nzp_4҅AM>K08a"Jr-J$\3IsDHZ6Ɇap=yv&Va<)IpO0k5f,`2b=6M."6xg\𕐽Mk~c^$s\<Iq6d!xXP$0dC,pNaU h@QxT>s>]`$1ƕbXX- 5^QC0, >Ō{b{Ԇ}T%^j<ˠ~6Va19.d/nr ҃.C{zZ(*m`-s JRIT#i,\QOaƃx`nh-R$|'rK%bdF/@ Ɯ 426&zdlVi :}1,o:|w@'2^Y,&O.W[\{b#ɂ3i=#,#%9uAg !THIr~3$E0@diWU.Zdxy,jSz2@K#xJV2b$ AHQĦ&jvduYCp:G,u%Y/pL}ڝqǶRw,AjG.(2]t`R+.:.wS"cV t*`%r +:J4$k uY|HY%ɨF.9amO,`l+EQ$A"xCi!U2WlH d@2%E-hf,Yo}YΘ6rMY|emLD.r&-6HlI/Ho8ЪбD쌜>4~||qɶrQv,A.*g$Ndc&ї%I: ,V` 1ޱXA.B.<;б<DؚCQAZGm؄Q4)ΣK,#/Yq=(sRNTڻXx_YR%I$R=q!wHV0$P0/![73TK8VO̹anjJv~A&0J=/Ȑl@A2u { DEP ȅ`Ic2X!ii uB!-q<'Z]0Jq&M("Q폲̥,(cE*f~\hk';Qt۞0o ,[>BάML͠YXWY*-B$ROB5GuPMRI=Ⳕ}3K/K01J dCTtpnhEZ>\0n/X31GHgr_ Q_/Bt hLɐ`uŬ>S"-[P5w%۟F5m2|N,FlT9kO '>fR{"-;+R+|XH.mƕ*'- 8#}B46@| 5RCʃ"yacQ(yXnOWY\3b^_v~䪽(^ Kn[~<{j|Ju1^OOG sKcoyÍ^:NM"5C/\R3.>̷Gܒ!P2㒉+vԓ2/^jrodqU.p@AȌťrtG֏rp뜏\}ɸ[ }Sx׻[ԟI:99:TITILT)y7jECN˄!=qV"JF0NֽyqџλƛɅKb0~/ _?5s+9GzI<:o68x d%o .ۆa=Da-M͢ŇD_WiUգN_mԶw5I6u\8]#e`E(Kaq̏xo8g/;_v't4{uF'G߿7woʅ=}oO߿{CO4I)`IQ8ooВ14thia'Bv-d\;6fqr$eI }w7_w ۙ$?-yD`l|p&~^?_jrJׯJw~ӥpt(5S;'!"? GO8?mj`ieQY42ȒGZ*V b^o꤉zIH ]kQtwHR0%![B'Mp6ٴy{3ZԲaZe^'KӯLM';?jzׄO37'4@?7s/#3ЌE{wdݗ oŏ:Qyz>p< p(-S1Y1i|2^.WYeg3y ­io-xCJfmR6e{P3#XsL^ u|Rs9I8K"WA;Frq(q9q+q{u_C26"oNY iQGK>xn9hn"K<2:H{z.)M2ES>uYj =9 [Q̹Ԇ+ 8ԑ|㸚$i~-2w}vUtLGGxqm>] κ(&*A&b)LA *gȃpnvbXp{W Q18<2mC9 /A`L)YVhg@A̰Rv -uJ 0Z*:"jAH <"ڃN Iuv"6ry*mYxdv(33n]Ise,0fnԿ> j0Y'_/zW#2Ъ݇ͥi-ޮ'yw i#*fmHLIqƒ 3_"igZFgO{{޲>[1mKtK=!/%=D16}| A.#(E²AsJ|0 $iR\Fަ.x$y{i2?%_h&30Irju;Ѕ>PC˫Ҝ CS`7RzDA#8郲# A*9bS9snL[0m%7)OXKHKw'`RdGѱ3r~e:ap|ЍjZq ,3lm_E˧Fcqz:TI=3jP26h%'ZAtMrSFkdұdN7C4y&n)>d͆A~*/K> c>]t\3s:3c@CRsjtHTHLJ - p>\T0 9rt?vTF 9 "sUP*YA LX/]uJ/gy`+e0wep ҚΙ+2Ypk%FtYY s:iij_<*޸2j ɴͣ 4=#*_PB0iP=.T{H{L{l,wu p:093CU6U]ʶL ̡s7dža8oOh>KL02ZeS^; "2 LVej9oi=" ,rHqh@.2KdrIT&[vFNU|0р&*xdfeTKM|eI$i5D"8YY]Uggɑ/ޯ5CWwzoj}BOJhqnoƣ(NTQ3pqQN%u՛#O1*UPA+ T܆b{߽_ 1~Merװ"(c=S1JԂqme=k >]w4 N8 I=5CAqB8;fe`k2S*j"c2ehPgvPvDȆKL{&`@G@"1LN e:-bYTAuuXph >_\fJcsf^Z=y.=}|SyI!gz9_!%'8Z 샳vb6ȋB_%\RGT+E$r5`Ꙫ뻶XCw2gwRd@"`]Q0EZWʮ"I1(k'ӕ\z]ë {ܫ yP,<>0*~{j82]#/cVqKLJOC"[;"ɃOR:qp }ԁW2$ $5ܣ=I/U`FjуRRhe*%cJ TQʐ(,ɥ-p{xJd=tv"&%LFDH[#{sN/l)Y+t*&DOb8R co!E!(b445qPW^9M=P*/U(Ql@-%U&jA`tb}4ulko,MkdR Rls$d~ BngdN$g5rgWBnc&ɢ1j6 9vIxs䨉&^\1JnCHmb=ҥ=?t*[#ԧ_((5Ӂ}:s1#eDҩ;W|Yh վXqq7G$O.iqB9drDvbHR190/t!lSH#RG|NP XWb-ljm(p,t,2DJs{tddx3aͼ4gH'|2&}]r}*o4ҭvЪ`G)X.яjA:bUx~0QށF/p7d|(2OF{^{۬K5u` dP7:ljSF$cFG1ӊN|s#!o,Rj&PZZ海*%>դԖ^A]܋Ճ^υ \=LgäE2p0)זؽ+R=\IJPt0pUUPJRdX\W$ L`up0U}UBW3J}*<J޻R•v9 bŋ&C*-ҾUk+ǧH_NZo걨wZ=9d S86 4"l M e UQI^Xpo1Y2[˳6okh޽NEHEmD_l٘8l 1&#H556ޡ_R#7Sۏ>O/>o2PI*[el9"Gw-zIgoo Z ځfGҺ'f i~ph\|B.Hm "4h4Z88(-#IuwM5Ͷ{A$ H$J /X>J+WNF[u)mWBL~7ޣssޥD'%1kDEѓ 5,5QXEat>bE5l-(PN30"Y6ʨʭfIYg 3q7|z_ñMgzcGu6N;:NQy *ږ*eJ& !"D,A2SV>)ॺSd%c$(%9R$ ωJsl^}a;g |ΰ3 ׻tXTXmzU= aў|pϾpix>~8d L.J"+lRZ 0`٬qIm!R!"`' *dMe1ZN(]±TV`v\ "uj+qG0s_P3EǨ-z=;F\QH]u?ӘO"RzȒ iB6YwKWE&h322֤ ,)j< >bD3z1v&x  Ǯ(;FD#bwi<=CPCjEv<:Gk !,@"OSDt$,j W, PRq&IbLcI/5*B5cDLYG'ܸ)ٙ슋c\T=.xƃɂ`䠯(~Qv ATY 'D 9 q)pq_3uxw'eך|Lv7x{ЖQ݌([X=HwO0jFyΠ|w$dcMjCj/,|/Ryg߂ؗeyy\t]N/l4d=7=3d RS Iȃd!%$UwV\ԁCK JlS!r!%յۙ8[V).ѴP~ 廙ƧWW<=e(-z$Sb JCTJѓF$)}6w?jپN))JY"UF9ABN+xW>[Lg jBQ%o7m_Q t1d*Q@ZE8S06% ixu"ĬzzmiC*1r5ESl]x 䌑D6;! :4ԭ!X<2%H pBD~s1 )A0RG4E*kehEiekǥ懫WaA=yZ fy8a꽵Zc_YJhO4',x[ba~%~}<gb8:W7'O!_ߟ74G^60cyx$; pc迿iۡjho1"гV]z~÷W.any\.)aߌҷX+>/ѲZ59">WQOXl6Kf8D+߮KeåTcj!䇺1Yfxi\o'mvGq?QXB"f՗Jx )r(͏6{:Lbolv;0OwQ$iN}ĻwMVe;/St@LY` `zVqtjLkw^DUy:7T }UaQҰIW7;RONzCJVŢJ6@k#E vzTm@!(Ǧ.Ѵ/؟eL"tB0PC-~) J:6("N 'zOYk66s* y@EZH)Q"&b nڙ85j?%mX0z $֩hy]᷋Zz ro,Oa,y ɾD.Ȋ36HTkH1YVҜElI! ݿNܿO|pT `>Eg o_*fwg}GcWCY _1cNb?8O԰c7jLF%dOrKrk5)@=zf3}MDLIK&QKaJvVl#^!Zݞ"xtHY>]o;rWc{-h^ hHi,Kzl'yj%KVRlʖ XޥCp3Lւ"8郕j_@)rMtB6G0pذVϜiXvPOaZ5I3wS-e`W9vБA] qiB"Ϊ`YKHD*ι 1k]<9i7xæCvKh1 aǭywppQC' оXꚇrGg|^fzwq؛a1P#@'zo$Ri{wax5 [1o;pvۂ[`#?}nkmlE~ &S-\KZqjhCe!BZ2*`*21˷2 us6x+Nwc 3BLT$uǬ+U}Ԍ)Q' f +Ab:ϱ&%T"Q2EĔ-~$%eKuPHUI$% 5Ƅļô҄'VW~8՚>%Gy#.;vU~n'd{ɏl"[ϑCAZrCk.QT;ѰL](P)NV kMQAVܪ:~(ZePZ",',(q˜&Bb91X XK)RYDNᬔ&L"&Nm`** @YCs9N x2區OǓmCa׃ݿͺQ6C*w3its۩Unv #ePXDQK'W41gcN l`LxP:RD8 8TPZ3Dr*RZd1r8 h4KmA{v}u9my-v,_Ȳ+JƑ 6'|-82qd*q8t>ɧ;Fje8g jp{DϞTY%IQ rhASxA^^ ns,1nr9(T:/L4!i3K%ͭxK Fqa,6$R.y/:71|R:%#O/Ǡ[*z7 ȏ٢n_;1ڏkd_M_Ko6F "IhQ:Fϕ#QARBǕw(!r'$$Ï"@P@AXHA[`1"h8O(((xIx)Zv?j BDB>*$ b Zs&qqrc<՞J!pp.H+FC=Nl{GJ+"LVl)»U޶gx[D><28md@r'>*`<9bVfN ҇煕\i[IL'xdv i=yka}T!=QcSjPK )qjO)"y2\wabB'c'\`V}U97Of׋uZCG > ~?0wY/y7v7(\N+LRE)fFEXxGG%I-S$Uкp\~v~Tu1/7>vuln'g0#Y^zӫd3 ?qvqxy2IR?`n5B)yǨ~'Qόy=$οwt7N{d8? :3Bxϱ8g^}C{+D"E%g1R)g ǛI-OetclZc˭dz;vQ83;ĂoF+RgkgՐy as s>`k4/+)G?$q [ԑUn8S]I qa9'yw4"?ȷ:NCH3Kk\?ox7jӃ/ĦYnsCtn*@ xn {n?עH 4}Q |Y.kҠPv4RniSQyQJ=Ar%\ ^[ߟ]I>9O9skM/< )N}vA #7flgg54ԴnPNLS"<2 yƻ\s {\os.`ILP$!|2AM 'oc֛9iE6ˎ-u;e]1+΀LD GywL!'-HNW-=3@Br@՛bT%n\JHm=%!jQKS>=" miIKj10 `XXHkWTp0xe6a,#2tQMf|*5Ê#k5Þ' . `HI0ũ ιU`uh<$B\@ F:i IV%QN8e%#O!n1r گڟ]Pme˃o3PKW'AǷ.KӞk"΄y!FW{z|DBӘPE˫DtJQ4~VBg<2v ':)蠝QΫ(,w^rQDKʔ(՜Gf.H<J$GF Z/}ph{}g>g( !84 yLQYh/W2f!ĭqh n۠v`nFS-!DM/^Vfl}BBˣ7wCb4g BpUMG7+)ѸBQ&A*qZ}X> TVKQc2dg<ڝnƣ>*xI ; "qc( Dx0HO[ DHF 2nWl[zT/6毗J{Qл e혳@ízI~w3͇pt7lδo;gw8 3KZ =PjET/ 'x.c .ťj^2 +uQSVT@l奷$A00TzU!BΚ VJgiނk:%*E`%TInsEz"j7ڋgʒ}C|7|Z*HUw@=CTUcF3\s&R1g*1՘ Ly֋[YV{* UI2*V 6*ՙ?hesj/>cÏCl(%sSք }9 GGLY$' [(TŢ :F#rCmH罥BD@2(4YL)FN項rA6G,O$oȸP4BD5pHb)=5i;힝-G?GԚ? _oj}ՒUtf:y?vr=~}VNu>T|Z7㋉Cf 1Jq"+'BI>| ף5:Dqh{4ii6~1 mOȦz_vc Ѯgpfs ?A;;jH2Y_PQt-J62ނw9}qO2tDfܽ4` %̳p93=]]VcC3g/NvGMMrlTSӅ n Z@{s g{ƒ-"WANg2ue/̉W۰?!ͳJѼ[W>u6d;=7.|J0v?O}<y?ӯ-Ϧ=7Ǫ65^דZoOwh~R8NI*q@59JK+VKäH#Ʒ&7ե3xeT% b!V rVT&r3.⢠yƪO9C A R@ J=+&>|CBm[OvFIm|p7L&%c= KEpI[w^W5cT/ЭSZcX^ʤ.Fw-uHW޶GUPl3̢4EjDI\R$KG潾eEXGBV&, JC) e ,n'm)yWαc{]e~O=zυ\{`1H_i9rո]]`qo4ě?>[i-9n ?"7SycAnATX_g-_ oXb-p(g>A$\\|*EwHd d'mE[߬%+F[LӑW8{p߱]W*!4RP#W$/o$=4r2,$`|H]'WOqx͎n_nǺkq{XW]oևZmǔ7j{p)[Akjo()Jll͋Ogl{CqG~A~/J}^>2uja$J-@R/rz䇛ltڐ.&ҽ}Jv`Y]ﵭ/ 93'\95x PE^\Զi8g$)S& ѵKhb {g("Խ|?ߠ~}wb;ޯw=y{h \)ER3v;b0kS(=6nx""Ni݃{5:Okm_>oߩq4n,̅;:u5[T_jV}nQ:5ߺު=0TFڗTQzўS11ɭ[{{ͳKpʛ-}8яy ͗OͽY~ۏ}|'>7( Х 8Wg˷ L?/j}(L3`_yj̼7p5Õ/p5a QzC WW_"}+Km}v(%C+ W8+~<\!JV1xpw^Nmq3K\]يO $z!]g@Y:&vt?ֽx%>Ƽ6}xO6goInv)ށ:9W\/~[ ?|}17)1Y_wj_"?lbWqOU}EiCc(j(/ eaf^5}yzWދ?<4J nVa9\凫JQfݴZe+%xS-^lj2Dm^'9oWF.Vʾ;~?wyxs߮.eq#jF8 {`J] )laO3d^c+! ؽi!zor3m\Jsy7Ofi$חxw6. e7띻7ZJq>-6<=7NAri ڊkMwl0)wK3RœDjy{n XMq8' 9JRy*Ucl$ c_#m̔=/w{[xdڧ⌱ـg2O#\Cv-eUa fSXjNjƬ1F,\ ^31Bd!8|Ep2^'/sǓ*6SAd*%BJ=D< 9j"鸹`z$8g:JIͶm1@n$[3_cΦ&dy`tsEd#5nj9P39|BF5y)i.$^35z),W#rcH #݄bG>T1K(AEFr[ ր`VYkp(! gWb:K+vn' L'L F*UKF&I\3X ŀxٳn<A i2n'/3Ȏa4Z`yU"m%~XY##Z( v`[5rܼꐩX%CKqzL Q"D|&ٴf+Q J(l/`,SLl32bGe&ڕ<ϻRF͠j`o]P?2-[@Cfkt7XY:0L;&r ` #JBi8()$S&!C% `J ׭`ɧaD+'JbLp6%pi,T R!Ъ#K& 4EOT HUWRPC0q{d5{URRx&$y̼cIf:Bb|P-J8Pޭl-@Nr$F=a\!Kj>6qCm 2Ό8=ۮ#hG /M3i J(No* z0uR;L'k! /q>;A<-~<O4Y؏]@10fwn{=P\ZXd>Q t5K1{ZmSU&:sxNPr h`!gXA(Bd5Bu !xQHYrŁ!^b m>:Tq z(X*`t]Opiۡ&[]+!I r!ӗu 1ڽ3/ ЋUuD! 5 t+^e`5aD"=V . 5h Dܱ!-P"t t *+ R4I[L*gjыabh| : 0Ad.ʀG"Hq3}X!uV[UiX cA:ng0(!c[9QSTy^aڹBb84FjRk4,ϭf:p$bPV,bpP_CzY4g)/Ǜ륥|ݦ=?>;ט+Sd@@9`v>ypl:imr0gM-2aͽuflQFh햡@,ۓX~vUFv8p0)!A/Q{T9P/ϐ 37y(":U'%\͝1}E:8TP'6:Sa֭zS Dc Th ŗ cwmN K݁uYr%E*ahH , Az*:jTiXf@ A|W=&ӣ/>pw>3ZfjsVz`[FΆV`39T;7\̈́K,"Ok_r#+ޢίG.&7/{oƍA-,.CJ^րPk͈Àe ҂3^ n ip ]PHdW՞k4=?1R= .4CC.Z*c10&ٕZIa6By}k \O* rg.H]!wԢ BIUq<qvs]jA$J/)Rܦx+ݯn7݁8}OIaM.o0n r(̱]vg{6{+{aߴy;ѽٽ 1$B $by'tN@J]@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': NuA~kǺN p7GBi0Ki@ҳ: ZɪH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': tN (9f)ႹhFKNXuFuH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': ib>9pq,;@N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R'8h%軟NQx{;^| D&O%UIc\BqI7. +¸ttܧG \xv_ ewQP\+< Gp۽nW6 rFyRSÁ+ b#BHW3Z1W3BHLWX]pވADv_ (69DYާaqaG צ}/(I!UbN}#3 <ߏn›W~ۣomn7m/zcё~ۏ}<Ԇ<˹Xdsd[ޒ!cZw/rkA6ݏi+gnއ~ ~:m? A.wb~9:;ŧw7:A_|<盖٧߅仯y s8|zv~ٻ߶$vݻ7 q#KQrjE;,tDFL"[ͯ.v L2^XJl6 ttvAA?yMy73)Ke>-V̀X7JfhQo~;8ry2f2>aV9)YyG6KQF)3ZdL0&5Dw##kب;kHE> eӮ@ϙZ +M-xV=2-*S2:{?ϒ^UY{¸ŗFY,f!Yn5HsJeʠ!`~u5f&顅yڅZ=BgX!|,Rt^:-g硫Pʆ?xGW=Xi"Up[*tP T\BW ju(E%ˡ+5m+̘n ]%R-GtPDJP"Y|%-tо\|J(J4i]`.Pk*q@]%vtut%)BMvh ]%]%6%A]]"])JhS"=W&vh%j3P.~t`~-JdHpYkVV/8 lC9Ty,PŁW*݀mMՙYk$eв3tJްt3qKOzu*n]aFFmUBIYGWHWc[DWq$ik*բ-th1MɎ.(Sƨ=t{Ӹ2tJ( &*J?n]%JtPκD✊dXRJpYk*]% utut%MUtk &6J;@\~w`i`ZCW m.@5tbJhiWJYWHW雛X>J0cxP4>VӴ%ҴR`O8aX唩Ťn)S?ƈJ*FwG8Jde$j̑uGɩi{Y:juďnx1)pȻ#-eQ*(aljSʹc>)x(gvβ;ibfnw}M_^~H--f(=6<,"+C:}>\Լ?0܂) ZzSJǾ_ǿaq%)_~*4*RV {_@f||M J~؆Gխ. CgM?~W>uU "ѧDX Zw~) Uy[{Ӻ&7ͳq3- i&xg9nuA܅4e8 g~ O Lfpw=v] $0]s10 B*: S ϳVc&y00 USLg"KP,Lo\kI$3Q:OiQ i |؈VO@`GzrMsr$I'E9ϋu1u *MAku0Ei?]OO4!3l|l E<3 k,C8b{;vag]qnznY+/&*+}PSr{oJ70w!VN!իm'x dLW0ğvr+FFnUPYxpҢpLf &z[6v- cu-ĥ\e5b=}^6#S!xipU }yX`dWq3CA7 *m9=:Wܦ ld/@~Z Lﳤ?rKeh'3~\8x{PZnWpxunUo(+ʿKZ1՝%tLb$wWj@F9s' ΠtT:[_=txuUݙ0{JfP2[J/j~r5WJxuhxJ恐깬1E}4G|ZwOya+tbSj73jL~I T7 azAVqn5옝$Lidx?7Ko$ӫ2X+,& O>eݓrNXxgpqw~pY3Ao[~A\y8`d s7vety`<+qV,@j߭3ղ{  fI.& _גkOɨTlXTk3~%0 EbK2ϦH}χ-J m;}DEz Rab'?G䁑6h,H#m~%_n vjf w\ύYofF~iYƜqjS#q/q*"4%Nu"AsI3̭p3Ra)H${nR#%#z]@*@A:s(joFI* #52(qAwZ+n'Ympζ8n80w=F/z]] Iaâ\>6=vpfY>]qBx-ȭ3 L Q8LDF 0VN:'\Q ƑԪ+HXG"\JlB OQc* ӥQFYJy>%K5f] Ӻ4Abci *PQ&RhJ$qǰ@tL:06 يBwYPzB+U :`z+%V5S*%dC o"X ,'U)YʿtI:=r!(X & ?LkR)V\$1炖"&<5HǦ´u}K'tgo{A]8p_8P\8JY=W;0-2 N?}MqM@΍ H, * )!N"(wX Q OP$, F p(S$SGh,uRm00Bq VNfS^aKkoF,B0#BXYL^jʈhA #( C)@th#>oܿ\Ca|CG*`pJH<#N" 1sdP;|a6^[w7~stqZ 'Tdbky sZdD)1lf%Z7x' TS,O-5(bPJ;%^TeqE8i ?fQq-eG)7j<[% Ȣ`;XhHY"Dbͣ vϙLx SAVGBP2Pfvk!Ē[cmI EDpC@G,s%0l\qKZ;oBm a|IZӏ?Չ,A1! 5xfH;?BwXMVxX>Ao>*͊6{oq0LovϬZvuDKYQsc mq,$JjĐS2EsB5/_%:k-_n^0NO3%8Pw]d3K PvDY*'2 ws)ێ^,/ŸAİL00 0Q` NCAQcЎtYp%m@EjlgnҐ`ɻ݌ߧVrxz݈Ɣ-v;#wX!L <Ѥā}hRh iu|OgQN[c0p(%V:FFDi(J'I'>xd;k`}u٩ l#337xI"h @0yaJ;rRQ4J)X"ڴD5FKe}hZC: d =(n.O\l}8Ț:H<)ЗU #8M*((zW@''},0JW>B~`w0M#x^XWpR_aJ|Non!-'A򦒧캺\_/z+h#dEgUG2Yjy٨܀SS:Mx>e*1#][oɱ+_;R߫ۀqnb#=y1Ř" R5^Ӏmrf8]Sh3p|zLN{<ܷ 7 9,-'|C#!'V mp%!>ƒ6\_@ m6l>9=8j}onMW6[ -.+m8&z驃ݟoA|S5ل Ww}B!׷#*5\Y~7 y=Ht )-$^i<$bc{Ӷזm<ِG2'MAMqvdrgV= l'16gbgLS <,4 n$tOdB܍>2I^zk&h[t]݌ n:6Y,{[v\yn(J0Q#6&}%C%}^Bȕ }JaKs{ s1B4Ue_VQ _wcARKA@{c(L<D02=d_a UjcdUz@~au)֗n'M--1vzqBR9h#1ٳFS s€m_j{!E/-h#9)Z@9ȓL!yR0ܩB񘼩[9 ΢xN;+.} J%.29rD=]Tvx{Uvbz,FׂLS&@{B%0 0'r#me10K &3LlK@'R#"$ea&hk^EAFfAMmqrKӴ`2?(R]OPx>8 שrw +]Z,yˍ lZ=[x_Zye[P!xCoA`fGmI3Nf<$1JFf4r{f51KNJ!fEm !EC∜kTmX͚RMVu45t*2iV9 sv'-g0O~wA|4d&}d.XdPq`(MVJȌyRX&).Y)R6JVfR=IEmJ(oʮ3!crI9C,F95];vڼ{`ﭣ|si'誊0BĘ9ynI'Ua0x9&ϐ+BA& ] F$a)BHOud@u9a YE1E#V]5^#5}$AT[1VJ);Nh$(1ƢE[+1q>.#(1d$ +ڠ*kjlֈ?g>:qɮzQV֋rz>gINde#yRpcER@ǢB{z/wUe}vӇ6l6z6rc4X46q$}e?j.E>K(b׿6]'߇&=b9j'.4J!6|l$$YI(lSwq;uR3؍֤Ĝ+v \ƣƕR_ 9r)HA^J}1%€X%'*HmAEYV޺|ɔHVm9[f)i2}+˕ahȤ&gӐ,.OoQ5T6dt\#YrGuѡR^3:;&,B@QIWxg0t eib`2V"+T:9D$$AHF"7oq*-CKugΧm_z.^T4'02D@ZykZk#Ald,@{]Þ?iF2%adN)Al+i$>P:"c 5TWv ,"+u2XRpzL%ȼ*T/M M2kTED0zrh ͔2k^bHz>k]eo4 8*QF)yX ~_׆H.fw.a~M[K4XqnȣGb 6.!z; 7i۵YȬJ@Ǹd#$ 8?8Ik[CHWq ws+^ ƓqWu%jrGiB1f$+ڑ{uÈØ\e7TOZ4J+>.=sx2p~TuF] dQg%2R>LCu):Ƀ6LyWt7Ӄ0<=&v_~oWG~ {/޼,h)Q8XG5&5__}ۡ%c4rhiCs]ŸB|5㾾c\ð2IYWvBػX/+ej"qO[\V97JG/2p[b۲1[A +(-{'a$};+H,"`c#3} MHdcɉOQks*i>z]d+w,I SP 8ΐCj!3j+KY*z~SeMW%֪Ď Lr5awNwܟsKn>mivw8i;Ep݉J <2 ɆIFL XܐFeQd k?QOkzDߣ&џh*xA)UP 8%%쒙OÁȤ5EcA24F0/٫ҦUD"&D)s"0-(VIS@%'!MޖY(a芪B[ ڤ ˀTGC'-cx9/A6F tȶvj|S ),.*cQr'Az jCǜc2!BA%^XCXΞ@hKRrxƣ-ɤ +B4Ĥ))` Ib=}L^hwYaCi[kN1Hp%fh{(0G)a[t<*yԡLzh_ u/?xYm+x#y7gbHN&Ӌ~t|7T=kgߟ^0j\jۈՃԜ]Ӝ䢥0IQ^$KEZ#uU,ճ]l7PK1RtWOD:w..^[mg.ߺI^M7DU:}xͫW|cw#}w3?,_LMdzN>/|=uCހ{ ^6N> KϳvV}aiC746?Fw\a9A$6tEpT'\웠ܱ`u+N+ ń?YN&a~# N KDK=x"`2JCGr:G'(s0e *dkӞ! n PBQN=͜Korgtu[ԑ~ìYi[$Wݲb6={0'8yx4(ݘX~ u$ޥy1:Q#B .FJΐsS:bA833ZžݦVlp>ƭZdկޮ-5e̓O\ۿݼ^;h-Z@K+Jr9NKMVy[Ӡlih'7kri~rl bӼ|]'r1Uw1O0i#ɸ~=?~Zkl T:E^5HgRػ{s~L9cs"r$a,pIReRR,0#sTuOJ gcA P:%>&2J\ T"HY\e9[O=t:x;q>rND~a7/.W !6x6?Yؼ̻y|=Dԥ v6Gm1ά(R̜ty-*T:zjK\ e-r$r\Nvs9WC<7]9}?p޹d ; Swv;'6(J-z0}ϡduꇴ-SWW# v){0vvq^}La>x{3 bR:PUƲYJNsYJžF)AFukT1QPV pSL4iaw ʕZUL/e]`\֥-j]Y@-NtJU 7ߝNSe˪ҮJVdS`0(V`MN$]WieQ{upWЉydz]!4.Pd?7!F&~݋Ϟ}|< ʲ,iimCU+&:I4V^L0UA_SմJ޽<;kMb߾$)F{sO|7ôo0fRX[FY9.u~c7QԩdO\+QCa0}uHfd!!iW+֓b=}ZBelR R)ẌQ }yW/嫹A76!_}Si)d^K=u)B>1!IajbjһI8v&T bLud&Ut* 5*ԴJq~v[⾃mn0Yc{{,5|   `ج \@Z)_AIi=hchlzHWb+6LjVuE']PW;,#]!T|+5օuE&]PWtE| l+|4+Tt5B]FrH ."ڠbRIW#ԕ6hHW>8|Wi (CR\7VZÙ`wI뢜\?~l6.7#3;RQٛ"[/WVx7g9\{|&m Jۧ;k#M#|J•l4M0EMY ?0ᚁEܺj(]l8|]mtEp+ 몡L4#]plt^x."Z([kxt)"`GW ]!GWD)!j2 1銀g++5C->G ),h|^R=>t&L8Qcp^݋VIRuE}.0KGWGWH{uE1*(p1{p=pUy 6v]%1ʄ☘f{d.t&v' pyzh6t;Z|<5 Mqϖy5lhW紿t:䭹Y%{<+`tv_Pxk<)Z]iMwO;$uWεt%Fno% x͋l[{|%k}}AIoӫNi(֮)]NߥRTLw(ln~UNQS.P@-rcxaD*\OˋeUIvNۊlRvnlrVMva*r$Ey(iNӪ +p,5vIkrخ~ ӵЏ2Уk!m^*c+2*EWDU캒*ئ$]=@HWd+ V]WDuuU 1ltE^pҚDvDIWϢ+sK`+EWD;-(uH85k>Z6"ZbRcԕ5V+N"`GWkpт]WD!jr&8IW4U$\˦hĮ+ "5Ǩ+lX˩ {õl+=N )Lƨ`hʹENl^ p?OIhz,Zfur/J}?6[q,lpz =g n!ޏ SO64V&]m)X=='f+u&v]e  ܳ 1KW;|bQ]QWk%,#]!|tE6pzQ&j` QltEhC]W{!]ue^2gĵ6ueR%]PW6(#]٠Qlt0~FDDJRt5F]a [g"⺠芚* ^aSuRӼ+OtE!A4->Jz&]Ylij2i 'B4f%}T!sL%M_,'AAgr~YS]U~.?]o0z5LS\?^z{ṱq"j[5S* Q{1q{oBvٹR.Otq袸7bTM푟9)Ję< C+m& _~:|Y'&;%RH hLW/5w蟫?ywt%uM ,Zܛӱ8 =n;1+RV?WvRcNx!?Tg5ދeL/8^_Pr۹*,2۩ ^ndw7w^lԟ.חُ?fTId3*ꈧȶi.?tyͧX*O>|]@ڣu~Uwl ?ï~{u^@Mn_NKz/\w=͏E[WVnuթW @d;[q=.F2/>Lw5n k3>a o-p3ł:X]]z['nPgd쫠{% >@ͲfSt=)RiuD7 a᭶cX(:mrpMf~C؋6ȁvG";*zhz HW{+tߒt,mRtZy6"\h}R!F+/Qcqb+5>v]Mu 銀]n`+m=j<} ']lt53r](!j&ǩ1R6"\\tE^9Ҥ1YJ29]WkDWD룏M<6 {>#+U]I.D;=r2O98!X0=6XX!|4M5MI1I#ҴFڧYed{,ΫfW8ǂ:o0݆#+K6?l~iE`imu\.@u߭G@8m +1ʫgJJn6T.uH&ܪ*g;տKΆY)g-M.w: V.:AC5{K'xjbt3Mlft&HWY1qp+"Į+1 @ ]!l+U&v]IW#ԕINAv7.BZD"ֶIWѕ ']pD\VuE>j2K1Hp ]WH 힖t,A c tltEh]#D jrA[ptN@p?(K|6=z;GW5]!"QƶQճ*XZiDUq12NWB7,B܏6 I/JiZдN601L]n\t^uE F+AHW,x/Z2CEWJB0Tb3Q?Z}c(mui/_Wf+6}WD]"JkF+KNFW$](L,vNCE+]WDiRt5F]9 FB` ]g8x"JH#[Y}Ѝ7=(?]TȎV׼Ⱦfb4j;iIk6@(. G|PO'e>B&@.JV+ޕP{U[:]OS4 _1ⲘD6o\__|-Y/l.tVK'IZhc*.+{[n}}s|}>޷~6KmF*5YUN5_=͛qOyG?6V,#G 57fmy9+5-|5yE&ޮ!N0mGcSm_a@bYU弼?c#6{2 uO}Y(Oki #+H; *1DYYZ3'}oy{ok$iìK-Et>i5_U W,rjk?81 QP9؉0JK@hY?{Wq- כt"`!y[Aڄ)Rĺ@(&H8v6 sE.Oש9`-k̶ʔ3=HB*9SAoXjug7vM}?QZla/+ <]hstXs)Q,1[+ya̅1!v-D&[c1P6T1:VްY 8NJ5j*ZLmr9;~j!W^ARҜ޷^ qzCd"w;dBI9A\13gIC>>>͌eB5CcNBܸ윽B1qfFʳRL=y>Qфb~R}e# j)bgtQ`[jalTRa v bHƮ9S-X89f|~GS#%C9*g@ĬЈ&qC6)60Eu ̵jp3& >q,# dW5fʘ{ϙV u.'3CLXC-E u}o)-e[rS %[I{mߢ̵`Af3e m6f%XjDQ[GKFSuΏЇ9 GMU|x۴/6ךKlz,搭& K.es#[cXc@P*X tfдh!Ѡ X@w&v &kJD[( [<`-r-xbn򀢣m>1h-K.85.nZ J =TjՇL%~Y`,S^q9 Y7eQ\Zn=Y{mGe7R`PY%`%Vhۚ 膽Z<={CEۆ h%Det((.qt2mm ,S<56}w EĆI\nZh  A ɔ DpP\{B MW4-U`Rwo&x,/8-h&`%Un%ٕ'P3z3|Wv e|8j cH[u^ʄ5Hd g`͚F}>6moF$*XYHtawLƃckY sbK@4`=*@An&:˓ p bVF SMtVo %Z zaBMgMygT),dD#>@PS0:pWh*uy$cc74ZED7+:Ϋ#lF;X>Dz\Bt@mU"z,[-ΆŷG6lM#FZMiB(.d3S> ~y@wa+t1PT*&P۶Ld%5BҵLŁZW!԰˓]@O`}MBWE {0z-4`2:چ]jy t#h[=%0A%b-s 5x4; cxc`Jrj@[3[#Vl#xe;&{k _6H j a)&_/]BlSmfiI&#Od,Lhcc.)E@M ,-.x W5M,pM?k ) d0 j#`QҦ]Wۖc)p ]]+ƈv(  z=jC [jw ,hKo}AX:G nh)/$LB"rQ:׀7ru6)I@d 5\ u"N7ntK aӦ2,JXQ4uVL3v-wC:{ BG\!tf1C4M1%PY+"izXP)ΓzfjňݸL}Aώ#ȍ!GڗPF ᳯ񯿻?}Ţ>'1|p#9N-Ӌ)6TyDzaxĥ k5x `%f$70׻^ :H--0gǺ.T:cUx hN.vn5ڴAA %!LFb2⌯(\H5'е%G^JOKEn={pWY@N^k (btnMàev—f@F\Y2čvXQ86:K(8qK9Rw$[ЭxzAp&>%6Km\U'.,P˩ՆXM:Iynz7Tp+n v KkbC\~A+{=F9`Tj^8[oΆ H ˲PuiOK .Q՝>9{s?pq6[L]R vA-}@g_y E-ξwή[{v+rx}ݐ^ܟ[墏ۋw/ "oᠷ8?-7~K603D򞿗coo#s#t{y>\w?goG7o.nߍ۳ 0֫۳tR~L݇X3sKLT[xGEn@@@IL?v$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I M&NS(Ɇ$צ$'GMb($&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I Mit@𶼫$簟$PJh7I ǟJ4 tIɴI MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&ଢ%8~@י$wǞA@'H@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 IJh G7/䨩p~<>\q1ݹv! @{ .0(&}p ()m 1'wCW^JFt%(N%a?R$e|xbyKy{¨?<{7B fs΍;^;{s}{0z_.OyΖ>&?^\j~vSz^_o ȝ8Y 7{/WGix92!gx}r~q+Kl釡Ж^Њ`5~g۾ke 8| PM~=3_ywՄ=;k|u{ySd.wz&պL;eKK?6S~=0뫛~P#_ZjI2>/?Ϫ^GZϪuji{䅫]R,#䞵.da,GҗA<n&/ r lkYu-_ʱYsuÒKD^2_Y}ƿ0X;> ot_Zx\4>|K OCQ\ZO—.=IcOt%~Jf"qNW ҕ|@Bv;+ΎwCW^J|t%(CT:Arl2ѕc2wCW7ڽЕGOWtutȆѕa7tNJ_!(] ]y濧kW8퇮7 ] Z%+]"] 3ЕuQW6I]"]EMӵ(; JPvut.p+Mf7JZt%(NrpFYVk;:5;f;дMGO@]U;M\Dc?o$-M_Óoi(u JW_mMC=8g{+AkñӕtJWHWVfvDWlM ] \BWkbxc;otЕѕi7tL+AJPzUWHW,#byؕ ] ܯ>z쎝UWHW>q2#>nͥvAR;Pfbb{3(ODߑ셮BBc+A ҕǀb!祱īgz hꐲ˟͕Y?Ia39 B*PlXQA׋>')a9u&l끸b4 w /!*{d¤,%Kɔ"Fh-g^孌 YB_3 9uRjdydMS1Acj("/0._cR:(BT.+`Y[bT/,Ng<3\o[o|u6!\X~鿦j8*_xᦃ5/eBu׮u:3Zt1'[uW4^螇nyh|0鹛\KL IqMmJ)@!\ɢJ6NN0ʬ\1 pc uSY:ͽ_wzviwKR>jCqfxDqPTuYݪ9׌3) 2P=S$ũYkXF 6JyY޲Αz_5ZMPХcT& ͣNZo 4g$qt,D W4jթ@*0?'g>N?'%%:IB0JI+ 20QD`jtIVEcP_P.2 Jt%rTn"ww$t7B1%oH4F*A|jF({0c;Qt6hvٍ)Lʑ[|2`Fb8փmv${ޱaٓoKz1t h$tdJS)j5C&dq@w#"lk^@|ȩg[V>T#ԃ'HnW>VK:(}6 BBcD4YНzk0Vj\5 9Q;8iW; kpK_v&EӤX6Y#Uewxf7~K3uoZ%_rPL5&AEF4YhqMCz[e> ̐Ε|dFa1#Kɍ|QIJ)_dgPn66iMgHM]#q-rwAh*ݹsJpGUύ\0ïrkw|{ՌvvߪzewS~kٳ U+}"dIzKҕky,VW[]vkn'fPF%Rv<غMsbv:G6n} AlY`b)Ǔ~By0eZ.qMy4Džk6f}͋BkF;z]۷5Wpo !EsUcCUbzۥ'+P a6*)\IeeS ܾo3!:(v=RMJFL둶սŒ-舋~TؘD! [:)^88IDZ%ϨKv_zWA Ʉ$L$a-lt@4,g\[/,  4%Y4AdYk^zvT2|ޭ|2OyKG%)RLpBMR!S$JMAZ4\XiibfO(4LjMeD.4X.qs.AE`j>jd$ZEmsXW@NSHS|J1O)0&M4V}r(BA "BP:h<)6[ ݇P$lWk$Q89Vg0m0&%N% $#[yP΋˓"oC9阌,q/eR! ?j5&@$SL!l\$TD9ӱ3ұ jsE~ֵ:w(n[A]8ZSw[Ksar)cmAwr=IYtfEjv5LH:`BTk=h,EN嫈&g:g!R|[LoRy4&P[o0}(G?n ȇ܍]ȇ܍[K̥ D!(rϖs&rizZبPNDWQN(s^ڼlRm * Gm犓\q/XB.H%O8f@SE&ƈ2p ?Ēx(նo%vGN}>1YSqO mG $"9H>B44 jDL(c\b$muTUQhH$ҘOYCs&Q8S1OYr"VѲa%^eY;0%˽z@hg~4/r2AwU+N\㿽>KZ;oBmjin|o?;?+aўݱf6qkX/NQXԆ~'ޟ?{M,(I€Ag[H򩨖 OK[Ee\`bCOJƨw/_%:kO|yߎ ̫vWAo'|N0*DX@ug)$m`se"@%2()6,ƒX,w$aZ  (3 )hc%Ёo= _+ѿ-~rMyH> fr5[9u?GycZ362PCbr:oAEƓ˫ 2I"XɅ( ~+'.6 O5bgz|ia=@JؔuOaN]`d|d8:Q3g"y J"#T*/rBơQIK%ZSHvZߚ֟.Qg^{tXp`SwEm{v?x&[KmFO^0q<0mFMsZR` DQRbW :7PuffKt - P/>oNxLU&šV?qʍ_6u=2zWNM/M9Se/Qeoہ&opnW傕A6ҫ+ sɴ 0?'_N/(!&06ޛkRrT{7/O/돯{H?v?ܙl=?o^7A_mvo _G8MNS Q? *l]#d)pgHBӧAAu};Q ߼ҼKۆYT :67m*yma%IECw!'g;qP;=y<>3RmӞO6v59k{ I iAyx ]l?'smpL>p${1.IwLF7^6ݜl2sڰ+;eS1+ɀLD>mv;ژ<Ƃ̸:0;*bT&mnn/8J|V}H}3TPE 4D= uThC,q a^y8h@ c9ЧWuƖuE: U_qvX_h7WF9MKygtmv؄m4K`." H#jj#'۪mR!QCpՋ_jʜAtRdjABГ ?G|t$ ʅ Se+AᣒeISc:A :g2t=Dy &g F:*#gM> 4 {{ ?%2  X͜È#'ۈ :Ȳ2VG2!y FA`U( IIjOF> ݇r28tڮy95LI^OMDd($J&WA*/WX'gJ|[]`kDkwCZ??)d~ X%h; 8)?hcNTՌJC !9GhtH$^K8l :%a!h)&TR5c1rkrJ1]X3ԅiJmUug iE;Ҝf5y}φo/p8o\c .5!&AnTۍGa%.jmZXkӽ{0{(a*mFy eHIePT&z%ERQPׄHbiGb]H2Auk bׇMP@YdX?ՈFd{׈q<<RDƣ `A +bLbZB!A[]+єc,|$,e=Z(=ƒ @F,FzltSlyazqx"'ݰڑ+FGѠbp4%Ψ,Kb#zqz/wlEa}(Ӈ;Pak6E6U,ŠFр7яO~蔒Džhd1Û_붓cAaBX[kphXpMuw dߐmGPڣ$;xvLY9s9P`']_Qhv93E^1 srpSt`:Cxv[FG*yDQN'RF5 m8sP͚T[cDZ&;u>^UQ"k{6_gg燫jN$؁*rqЄM.NfDڍn<ꪽ~o_@YhH$(yw˿m:4'D[ ͵`h*K>|71{q`v& s/Fgb}Y -V#Av*,~2/WDyoqt* ,gQ! ӥNxh_ DG mbnz#ZBbF2spB MsȘ4n#}dBՍDŽC7Dn 1pX -ES \Wk:EtzUbOl `bK1}mWvza0~Cg=n4L爁a׉BBxM8Z TF%%0~"OkzmEN<-\Uj]qZ!*Q >0AiU # WԦUdZ@c\ i$IcpRG3@b)z(qoIhc ׄ(t4#t:he t:|\C ),&w( N ² a6G=;S.%#ar0|PȤ1 ¢:`2tqrCiu>+ }K}#똡!@zf{z@ebn 8Oq0z+ZIo~]u?xYm˗<ӷ bƓ ~3q0}K{1#W/8QJcnx35QItg( ep,FYr,s䯞g?S]x/Hl_Eԇ.F'$t~ZNrZ~XUnX4^>Q=yۑ/8[7h*#zu=h7T ^=* ]A8 }SFM'  >PK|'T`9c؉VպƠK.j[;%08e&(,ld;jc'+C%/SĘXp)4CaR4Ҩit0N J>z"){s /UVOOy/Ye.R-QqZ(cƴxGa`OF8KE8c0Ђ &sjaB):w)x&{h[Ϩ{m {wno[r0;.FQ{iw}q,<ϵ 쵃Ro$t)RKXiCox.O㨚ub^ET_$7Z{;wrԑ>δzOswFy>/%X\كNp?p eV~_+9V 뮱{ZOialkY=B*0c?H?w"WsdkQwώqwv+ZvS=ӚʠbZPt*XRsrQ2EDm~$%eK빱1J.& HJQ9++F<jS"Noˉrp:/|$dOoAk []cb>U<ML.^x:l<˳J{4Jy];DMW7*w?gOHQ:[˃VeAa!C. Ci"'Rk/DLjM<Ⴉ2+XGMv?!v~⬯'W߈|20'c7~u| \T—!XΆ -W.E+\4NDgfe{CerY?eEHo5Hu)ǻ)Z - Utb䬗{o@',uiYjk=>:W#',Aj(CQu[6orծcx[_N]|5s"8UkϵP[ dB&![)iBHI{BZ&J R0x"DJ'Lq*)1snT0'@/k6@7|j;0aF>`'Eb4:!aw`ݼw16&$Sޭ<73,SxK2]d>~y[1#cЁ#lu?tN6Y5wL[@X51䰦|{f}q$Mp#eW+Wwҵ*뉦tFRyv֥bu\.Nr_կUb4C7HzoOE#ն,+lLf4K"7kBKڇaRvnIC?<Jk}4TuRm|n FLzby[Y}az\Oot{v3Nrg>HqO`gӻ5Zb&31p"R3Ip=i$>iId Nf"#"]1p2z8\Jh)+,c4*ީѕ5+MҚTt5A]=jҕ]Wբ+u6w] %@u`tDzzWԜwϴd߻JGEWU{Mo+v'.^ͺa)w]1%آI#bpf&IU=sMo$t9 ÓNw$8o|pZ]yʅsZ:wxl9R͹Oo㚷 xylK~dKM#0 ;QqcKB  =^z)s] -e+ EWSԕC,(ҕFW]1mt&w] %Ģ [kJ}T+QY҅ 1)w65b\BKJ(]MPWD ){E+knp#] m>wŔ ʓ7QS0JpӺsWB%%OHW U+ujrWBK2]MQWуqzWJp#jZ}J({-Vtt<0ΎB:6sȤ4-׎DtmAPBbSW'K/e\gl~?>oK;Xuv &ƞ>?=7f o&fVzDwR9?_ՌcV.5 5> F k~fKsS:*Ӊy8İY?hO6I/(5v j:oTʩg): }q&C*lV8PY]ٶ!4]H6)N H8wӗlqґ ƃ=FrKhH!0ʘY&4 S _Z6 Q8Q+%5rSF&+*ҕ'R+&hѕzEޮEWD)(=а-]1nEWL]WB}u1%Ф+N^0GWB )w] e̋D0t ] Wu%䮦+/N#ѕFEWL;,(jjd` zF7-bZ٧څD)*HI`2h`PpA6 B޴EWѕ'R+P=c.Yr9dGWc-A[y׸8&iG5n(e09Mu5}wmKޚd0aۖ7:":וP]MPWO 銁c5b4i+$[t5A]4 lMR+5+2`pB`wjt%1iӒ>J `銁5jt%hѕB]WBIj)d` zFԌ m>w%)]MPW"]10jt%AM0ut 2p=+3Jh]WB|u 8@Eb` =+EpZt%rוPPt5E]yzΓ2o1#Mpf9&Co&Qw|򦷶l,k,8ihXh莡{rR/).d_\^[YOO_v]GG<9;e,n>;۪ʤ):O)YIk!ÃO_ハ$7+``uv f o$"Y179!$ {63w_Ï9g`:!}1:tt5/l)0 u1XUü[xֱWяQYqiiqsVM0{Ħo G%/?vẁv/]/Ur`np"Hibbи踴n~n>,VA?q]ZمKǮ7M_$vh8%"W?r<]1?g'˟eqyUWYLu盪Ŷ7J\):@}Y<\;._;X_!)dLoa?Z ^.ϝP9L}Eu}rq^~]< *=ZuN'U4WK{JO362\?z)`BHj|drH I_$ Ç!~8Hqpr#=D Rh<9Ċ]6TP\aՠkK?RnY`p*##z6="gc$]UKi`y OW%%EԿb,;H< lqfhM5>@\\ܠ&?%.gW%8k/d)J$rfV P.K2OQ -BRѻk4d{)HlJck%j{J7nʳZ*mQgET:)hLo of>["NggdMu].N%I|CYx2M,DcI_ͤI1&N$ol x}"~d 'ݤ >F0EWdiʧ geu=;ElCrXs̋IGS y ǹB1b=H)-{4 |Ә;ǧ3u3]T9rt}ȽGJB q Bp_>j8_~,=~GӛXƳ1 q!Mkie`<&LGkHD*}SastGc:fDԁW#URDVW߆WI`*,&d40x1O3O0BX@<ձr\Kx@dDhTpZ zc`dƄ>C#*y#+Ø΁y R *#y虔pLm8 Օ~²w`K2DvAJUIݵ}](Ugv8ޙtk+mٝ4K5R qj1i4`yr;nkql~Ķˏ_u9ۃ=Je,$R(ϴe}Dci[dfGভ>Tj4ϳSz|?:( 4ٰL\pQ&.2: s{rgw v r:S8c6w~vK-;;Fp:x~,N&̻&-= J:5p5~N' m8}VPi n$L,ݸ(F::9x#9L zv0bHUZ}1b>NJ#;3Ũd|6#)ZGCp@>6'02K ,2:ME%bCF7ug:HݔL);r \)Zx_,8]t^p[)Y4%RQ q(DPRE^!DR< NDrO"y93MbVkG xт9Y,T/#z!/8KX{4I~.hJȂSX`KG8.>N\i}ǺYjE;kX[wsFLi5T)E&c@%?9L<+ PF2Ҝ>#,ӣR^.99MjrOjDerūX$J+3SdNlġンmYM 0J~ ǝّs}Ѯ7(iu19\~bO;wR'~E{ZI$U]mE{&B-:R ]*9RXԝUn@(.L ./X+,}.Vͧl4H%H„$A"LH, Yk7Cg ik.;MP>du9#[CTnUҦF^' .ծ ޥ l$.]ϼKWF0f^9Bwۥ+-/G3JRz PzU5\jK7q܇A<?#zn}3!C@HaͲAY]q V'EK0ޙn֫pƻ3i':㖎\} H|NY-;a١s.`+d j2nnKRLH.ˍHt2.4".9w; D1uE ~bdKR NTZ_`R;~~>V5[hSzB,eѝ|0@]ÛXD >&_Z;ɒ]W`&n&Kqo]NJfQ K_-Cp@@,՜Z6w#9l@W׳~}'wfzq3lM<3bYy%)znjYuH9R8Ml+vht亃ev\fLqO乸(qP,VVu',JdRz&nԲr)5/ oybYEٹ$R"ԑv*vɐC<.pcJa+b,8Ba4. #92JBdprݺ\ȮZkK\RPZdK7:)Jn,jJ%oJ= 0'#ʤ-gopf2qN9b%IO'˝3)``a{hZ3x K[OK}ŵl74ל. (v fj۔ ЉHo|ÚԴ 7f/EJ"UPDWG70%i<0Q)7c통7T 1R$x'Lg@^x( Fz(3ʽ䗍s .>`,"uE5p%Qs&?BAL5=1,c'XaRf๟Zfޤ1&*T@VByt4Ԛb{H5_|I'fuRӊ/D|6,~||Ţės2x||mx_]·eq,B2#FV SiDR]sy`K$|*E VjNf+`:fWj_RK3;}?bBRTmOz3g*!äOQ MQ!*S` ETdIg3* {6B=L!~:jJ>k9oڔ*.*L2"ʄHv:^H}$ ;e<<`IZ91P/J )iE]CkH1<7hlLcQ"d,/R/N˫^z?;CInc e X_ E /`h].04zS/8Q 1%P`"~͙|V 9WqbtP#$'* 4AKb_^QTiÍlP˖UK;s#yłI+fP^,WkMYx(K*풖 ]Aue5*}`Pjbr7:nm5:(kYBֵ@3-c;jJ (T{tшįq0هhVwq~kNǚ$[r+57Ix1g╋;tU@$P+ B[U~O@S&'9A+ީ_x\>~;qq w CtLN`&UHQEq/i`R*aR̴~6A+!H{ 8p5(hٔ-d~!jbm M^X-ԣ?o[CrkRi?pcv^-Y}AJh'Jxgz=]!*e}1#V6.]NWI!)+/iE-MLUb_ /b;WAc0u? (C8 ^(@?ÞZ$7e&`2 h/L 4WV:{{*ڡi2 qF%ljT@y&ȴP@!Zdwer&A/qr0-U ). QJkiM.rETGpiff Z^=eT0.<49 dLOAzddh,GTX٠u`mjDA-<"hu\WS v#a͑Bᔁ"۹NgYi|Ya G>S,Bi"-t"?x<4!(bl 5CǔQPY=XNɤ$$b*ֽ2=('D+eūq)+-2'λmpqJl9|싟e%HP+1@疃2 M>=cj!4όJh|<# %2y%{ <+OA2MQuJ"Ir2@X,AIL4%-0d`;7"GUt{61^Y- ~SPx St µ2vejdcXO_TwsPM!@-$toQ!z|-ڍ[:ļPBI?v;.xVz.YC[6g,γ<r7Yqw-;?Σ)0ǀ\J@flq kaJvBKb?p=0ҽO^!p i5Ǹ҄ỢQ'8*ɸS17L(洤tJ=@Q"Xc\DBH<ÈQuRZUz;}%MmX]xOcY:%M3hCe6M;Jq=;m&v,WXʗP\er0؀tTq~I}1n!Jy;Qڵnn?[8/B]ŒADI%ʆKyb۷LTY|ٮ4tKfOZxsxFgWܧF]}8W T'0wEu@m/Ҵ"/k&l1\OY*݋o ^62 P<5Y Rʼn;[0ZF_'n*kPK7A1?+ɋFFK Ʃ9Pw ޥp"/k (m-k:)|غ IԢ8"{wٓmí^乕5[zva"ӈ!H2mޥpΰ%!Ӭew`L>/\3LuEexLuU(Ue(Ῥ$9OUͅfJY &{H=?' rScߜFfPAmAU!7c@Σ9Y' ,@k BriX/. 3A®5𼽴U#[Jc(Anq\ ۑ`#V5S%͸4 Ȁ4qޟqwE4|dUSL"Ԣu P`-A! F-`u{0@{j\+GQz45j-*:ct4|0}}SEcF&-!b|S/8i۝vD)܉N^q˾HHug3v4Ŷqh( 0M(T#:sY*&/*2JaUP\A߂O}S0/rm`JO3B1ZB0>ۤHAut?K PV |ݜNg . 6.H4SItKKU\/:~e.!3H*q1j!T,8_~m\Ng_׹RW]I_Xf^'5_ސ[]+P̧/S]2|/'zAԌ)ܗx\m"*hh %h]ii[cRȺޝN?>Җ*M/߲Mm y&{j6ga ^5nٿC2xv__UVG}?oj:xOW/s8?Yz>!A4ws3~6`R*nu%Lg1/> "$f"{9/Ι** k&C*p.**eE.0K"u`X]3Q#M8+yzmWĻ+eД !gIaOʗΞ3cׯtP0QȣK(g?brécQLS,N)kmMA 9|P$ xi^B 3;s9Sfҙ#< L{c]fiIΓ̃^bI#od@!TV.YfFeE׮”E.xdA4Bיq]>{ZLmuUw5`"0|VxP~`>t|Z~̂J!U̐xfIyK=:( ĩ_%S 4{ Nm{Y[0٭\cTP"U~j!&Ř<qOP4sOv6og>3XR = XD93hK2a:%vhz HeKJť}*%XO aԜms* `,PUln5G\Q|]S{o` HXwi97zZT%7Rѳ5膉6ty)}G=nFJf 5n=-֞jbE{&A J=_ %{JRWvJIyӳ 환=~TRT]mnd 5Mst47;[@geLC"zUj52J C.T\K;H`خ0MtPVZi0IZ/C01^.2a+\#Þ3F{Ly|uƱ;&YܱYS~uB#[ۯ UYݼ0n+O<|7hD+)c%d(N A=&DՆM'8KHu G>`A1 +ϭBh+񞷪JWu|q +Db@470fЄA`;x7 *IZCm& $U! :Zq +c qTR&qO֢:$K^8$s"i݅fL[ϓ٨A OF]N(j1,ċ]Q<)i+R [28e3dy*m yq`8@ɪb1k\Yc &7`M-~K!aK{F2Ykz97(152:sX26びe2Tvme'Xc^m523&n,VW0Vލ$n&7CZ?TP'Q#[=<JԳ^bN'򟾁&aհ]  y !d`^arrH! "U*C$`tO"~ցt=]lsj612=(Y3J~tE7G=#MBl/IDy,e"qd%dufOxR}l^D 7/([H.VɗR_ʎ|ajLP{ϛ޳萤 |b?g04-Rm#ukmSo3lǪBz*}A5*FlX-ZpXd|A:^AiKһY }"2Kx 7M#X)X[3*XԣC^sr3A}ڐd,nUTzo(o?kGdP9#Zd94;[8+@7|l쭔BwmX_f,3 w"ɢ-[I>FKr(%n۲z{||ߣ)V"FφA,I!^;sAz8JO̤.#yiKGcB_;r+>!í'/ _D;ypoS^hՖ: |х#̝[>k;E7m\b<σA506Nf Dy`l nf8M`JTvsÈ}׺aRV@\0;m 2%;Ȍ&#$C_ޑ5#sSN3s`H!goaIՓ"a=II_cYMX 37MFYd*VꑽV9.EDG7S R:<歭1o6A8$*DhX{s=1m٧8 `E´nփO#F%r뼅ڀGBi>X85X(ayaۿ ֌ lC¢u-]Eu&aY9ZoF<\Yϲ8Ȇ".X~)*"!„.]DRKdfUOٿzbhqd8L17-^ѰѬJdޚ;we"3 iIEXefx%0;OUNɌ#at^ f?ϻb6*Oր:6k ?5 WZc$N[qybג+FW{HXFe4v>A:%Ajp⒇2(g8?F7ML13mrRfB*RGkl-"Z)QX A/*G6BJPU_7oROG'ާӃ ?:\ _CIRTX/%fPj )+ 4q+K'4pVsGX繠}}} oz?W ha9<+fw n2u>l'x:yѭ}Ϯz}LxAg ѱ*S)jX`= յk_ToJKp*91g7;44425eqՎA%}k>"Pw+Ij{2/FMuv#o~yLbV62c=ٮ2-*^_;ϼV$L#I2 XH&GoJU0iޕ!1[j2zѬ7G{N+^)M?(|/fEÙ\3RGCܦcO/_RMˁ\.x[Ί0}x运g6~à۟}daR.i_t` HQAZyx\f$|.C@ȫ/fmy,ωdGY M zoELh'ѿ&$0ËBD3v U$Fɤ/ eYdݟz!jPޅĉ?W57?3UCAKH + _F%)-uɒ!UBBBL3h|[ƕÓ}É7K %"T/$BH>0TنzK%P^ʯd";yzֽrT ?k&4=waxٟk-Lб숔7s9Ia~WᷛE,dY{FF+Cs2 M}n IՀoB7-;uW1N-°nr6x2H}۔ k(ێXr&*\d7oO2TP6Q XgXf l, G#Irצ+yVAJHX NdF4{e fwP|ndZ sx%;̛ nK0Mi0F?o ^g}~x0p +y^x|z ٻ(#Y q&- A Pj9%0P0@ eq#Oڰ=`]v?Ä ,&o z) \8az{0XcFR+*c*J\Q#9_>jEMT0;L^LV sT[ΕWem3}_mG}^A @F1+-Mz,Bked9Dj \IPtkeJ-sv2w+޷352j>#!?~^/i< +ƀ{P8)-F'K+a8۔Qǜ`~ZG~!rC97 RH B7WfKR)* A^A^A=!^Ԉy۠ܖޚNA!`F,.wjhlxBY5>^_V`jßôUdJ=]OU*[q A: %5Cjud%*PV|74JmQx~i,o8H=0%R,@3/U''roX۹MTƼac`E VŋԎGA*PWe)(p}9!cacTk[X0/"Jh(%=Zk@NǢZChf/YӸBEyƻ"P܇zxH, P^ 1 c6 %54VR'wz=?٫ǪXEP ݾ@-A(4wI-ttBֻ XI]I=n4ka[H7'倘U5=;G#Z(IʥkָLe p1,Yn3XVH*3Z['%eXnl9]sdJjc/oW3 ֪ci{9??f]ڶU^tj:-J2HF»;]z%m7&<iY$Cy5cH(z b*BK-6^AbZͼ]tP#N-t`bp5py~[tҐ"mSy T Gຌ$4bJ ̿R!m)w,i璇gȵQ7Vbh\coF.Al㢘{{=.ܨγVog7{ :LY݆"?eo?JNIB?m/{av-xکWu!^sX^BNĝO?{ܶ J/T/CJٳcél E"ة!) q@b@D^"4=ӗ/쬍gC;n "BUDi*t"O7J0y0A=fE?aoGfl6]rxn/NOxni<T'ylv,JXG0#x`˷33 /E{(IpŘht c| $QHnV>ջi*a2OoShNznrS )9H]E_V8mS6}](8aTQ\)H؆ŧB֘0#*SJ 42L %V4^˜G7S[\J/ 1NI,I1v@վ/u~d+D 383`*=TbH{{F,ST$qpq.t#<%AUgGtI `d:Id-6IeY,A6wBBnac)ۍ`mmn^`|C: pڳ2,`w5Ƹ;5Lq^M@0|xcAmk*mGYXk,RU[Vv, !{NW {ݏvy;阕˰3猑ـdNMCW.Wt GP ԪF|`GlslƨUbnH9Ԝc"!yEOߎӑi8f_IA<O3ԥˋϣ OxV8ƇSpm.G E'AZ5/M^PqR jnBprf|sgBQV$k:{auz*oF7u_qx ?[t:L 'B"$/s- Wu;;@C+h>>d%cN=dGti ȯMl.jbSQ!t,=;%7Wi6s?N]+h[_D[f(̾MUν xGofmrfK%0k}2UHѡ= !f;33Pӷr5یt^Ig9mʯIu8:Q}ƙjYNt2˛Ϩlޑ3#ڌ֍D9uV-q=UAT\}a[ى#FsG^j7sZ": ,ZD0/TP)^#44G22;ӷqd|ϟ?^,1V]<ݿB+M3R@_?v~6X9T |ߜr ȄI5k#L5Aw#mX}䖭BL 껎{'0 &đNvt%p)+ʌcLm_y(lه+pDzWX^q̞8s^<0qr̍ 3^o{9ݷ?4& hk%9X^b%ըl'I~iS8łq2 S,<1qIiq8+72(+7*O6m&"$}n(:*_Øzw}u~!L`'y0􆎎`D#ѽKz^!yL_+UKzva6m>,AX,J  M-rd(ô wXTۍN_o@/5V-m}%J #ž<۾lb7P_k_#d^_87UW? ;Dx5׹u<龈aylVu8FWd-.-xp^5N|gs/5-A (q>n{>muW[Q{ IFi& :T\$\bH5UσXd"ND(N*\P TQWi~pFG|,:^T͐\ (%&zz$%]84rI`BaBaFy!5trsқLJLonWkvx.%b,!a$q4TIHJ5"1IXE9l}k50[!':o]h'֐oWSiws)VO6=sW1oe~`C3inX۷ސY΢l؉b-t8iӧuK WA_HY㶬Rp? pJT\ESV~$.% Y<Cˈ{ZO9ruq֍ H"*ht@sFSW\8Ute/?hy@?Xt;)Y9>ml1/m oY - }nP4G| kGUVq _]pٴ'pz?ύ6&8EϷm>YQ$G4<,dD 6.7?@zA>Y(* ]HEȕPU8^H7%IB:7aMٯhӉ~nZV~ZX}~S.OOegE`mFKXLs+D)'-'o|5Ƚ][nsnD<j\OlDyhyɳ"8k؏9;~^Lz&/9{}N*B1o< <| Gw ԗn{EojQtlpOrv@NbHя(ƬƆFR8|O/SL,s_pw>\'wYߔozšn%GpGF&2 2R3K!Zd⧿.,͇MUZf s[{e8 ""DKxc,8ON}p'9'.o^pn? >fePܐ? X=WqO^-{ᑧ}8zNyl01~Qs{Ol{u2G A~eS0ɂۙ{բݵxYNWY' i}`yy_w@H8/$o1y`NujMHx3zu |ïk9пdaSQ\ϏM% ]m7>O*(/Bv[4]HbD"~g&Q"[r6j-;'/ z}M>Sx.OBK4>< }"KŽUP~ѰM|5\7eDs}SWafNu2UNIIEG鐯 -GsI.FAN?A#T1)XL @l{O'i?xQw>]<ݿ' 'Rp`~}o{ 7h]"jò{9eln8ET8Bدco~^n6%B-a0ץowm 7|ʄPqdBMvHjQ6DDve! ߝq&eX]UWbXob] b,umǾlهĠ.\»:qu-3{y:03ȕ4~ӄڕSo40e++m6o? ]RðZRId.SzM1b}2} i IWU81:25D=\xWw /ȣ$݈(.ϪNXE٢dI!.%x}p_'\(FrcF2Ŋ+K#[Q&<`{QL+7Bk .{<͹}'m&r4N|K9>{uf8X~Cȵb,ZA(R ;G)"+w9V8CFȳ@tZC .( +2 XD=jHP+ .EZYh@XD^n'hBzJZ)oK߷בHT_ZA MӸ)J/;k``*+ht<#.l"ほ00 R 528&Q)5EXA)s;oa @ .3 !:E!.,轭Cvi"ʻg6%WwrRHBf*U-lȐhN0c(ӫZ1C/+KSq >H$Tyhr)hNï?^F,*lrk񦏖g V&Z_,y/../2>-2< ÅˌI*M&iLI^煝 qY|ȮXsXdgIjt2a1qc .%} #(:)M"ah4U $6&1F%L 2!)JL"b$Um'vaa m%Jw`ni* :2I0xTZKm'*5ʅB)$܀U2)VTi\LMLދ p gzH_A $?y8ҬfFtCCU9%LƔbth &+J5cTCW@:%/8%"tUY+(Ք*Z=b)+elЦ!"qtb WU:,->'\tȲQLlflQ46F赭h|U.5YUC+R9;1~!!Ig'5I!6uoK-8q5w͑u钎=a6MwhwEnyGV \!V\ Q_ɆIҭx6IwS聴VǘN$C!>Ԅw;NR^XM%fvNT!A +)Z8Tig[Tڬc; yHg/{lxAR@OUQ6y7]蔎g*6j5 7VuSm`4m 57`xy^N//./]..ΦojF:ۣ&G/8L.E'%]ѨyS1ԝ.? 7f[Ikr15Ofp;w6 v]v`y|y(ˉ |n {Ǯ/AZНϖM4Jѻʩ|%/|FfE$AUE3ւ\ AoD}_ ԀY m5,SꝰBrWJ"c4@1dk.*vbP&pQ(gRG-Ca2v 5WyVZ^ " {APQVoYFSr6'Ʌ*8LdR N#U|}RAjnW&k,6@ r=1_DGWb57f02fw3LW)b\v^6TuOܷ/!oYA+xjx}VW6+ȑsO/GH"vͧp{yY-EN%ʐ֭|*೮!Zm` (&#՛J‰16(* 4(?gg|`,źR| *"hl)YI~dN39P "ˀoN{v|wZM~ Ԧkv~+l7=f~W#6`2z2]*zr;WM hcM%gK.sT6B QXE@xUasYewl\١R/UJ\+H"/e6PHmME"b8eI0,g}*R7:$.[ɛUVBI=Jf.B4 2g!Ȗ3Ȧ%{Cqh9y d)e e[Ý>EQgA}E[Uh`EBEc&!62yl ~?lIl_qЂto(΍_ݫ'ӣr=pgtVׁUNa^Ly]9{msb_Kdu q4ɠVS 0h!(>QO9$čt]@3Wy(Pg`z9]uF떵6f-Fز ImhBl"uAc263}Az8]3־ICq_Lӣm _ߥ-9 l:`8a~wYON&!O;C}OjNam1ʽ:5\5Wy6|aug&:3[ֿpQ8Ȩ?/C-8+g4zo1cy_6V(""\+%i%tQ HOk zZbDn_>g$L/6|;spyr"׷aa} 60.]s!{z唷$ϸi/}؉&ŘS}~L tkB{~ȒVx Y< XOrzƶ=\+vYj'9,67՗iI3-+ hؼef4ϡYNk@;G)}> Alzpզj, e3vWt绝pÝ?a5 ۍƉִ;Y|C˿kRڛͿ.NDco^vR8iN^Cpa۹R)oΏO)Hde* T@^C ElX-R9Ne\۩ަ_\NZKy̞]̞9:8;΃`Uڮ >f-FL \Q"bk0 8җo 1oyUP,VHޘV7d+9ZWƶbJ7x+u@g!4~5JjU1T8oˡ{(oX)kVbX>'MsX32ތ%̜v척7\K$r}[а̝ k٤U2v~JRּX#1 #yM<QA#!f'"|9Yoyp$õTPT.',|VE1&VlŚ3TJ愢dZ, &Sa QWSR>HW\JntG PCC+.iݸ}o+kF+a_A$}r; čQ.;ms|k%7/x1Y a[:3)uE1< ^6zKo֚wxZJrv~)z}GpHAc 寔lB[h0åշ*;bK+7IǢkT1q{+1Q{@js v"6!Nm2JJ%J[!2iO @x%HΊY-0$H&*TLK. )\uC)hw¼ K!ctU"j~wVbj|ݚv`%'\)!w^lA8PPj&]3|/. 3mX-ٰxzZ]rsC&~U1zJFvkWifx닷 !\7o d)HpTX\H$Sք$#"TmeXB%Z*C6du F 1Wp-?c)1b̳D뫓Xf2ϗgKl$whPЇ?|f2[:;Wޕ5Ǒ迢#a{cgbe-ǒKW*Kʪ!g?tRV >]ƺ0R暈X`S|,h;rͫ!:H`ձ7 +bCṷc *1ZMED /w}b/h}®CcĂ:Wʼn~:eJRMoc5x1 A& z+ጫvXVJ [Xw  ٰcU]B [*f[kS(JY$u/Y!7%c-|JɋvkCGQ`گz)<&Mٿ25di>qMOk2Wn*_MSǬCU&X$fkmE8tOED&@V}HFQR$;6ᘡ瑖t#"Bf^|?m;C_0Pl;y$7Q;A#=DJ]ηty'V,lV˅#*^AZ)WC-l?Q.MW8|e̾*9W濪7USrg%Wͅv]uW5W?(@fuD lVI*JapS+ɖdp(} dE깒IcƤRinZ7}Dzl L`Bs&טtv]SokV 5l_IF9RF)nd$xζL@ YbNkQggu3* Kdas굀WQ:K59?&P*ֻIZ ,o8KFQ@L+Pu, iw9y &;= Qn^ xٸE-7nQq6nBٹNfQIr~` J<8'^*G~g4u3!f~&$~A1&[>('6T` I$kkT,r $>ϥB㬼jx jKD:r8Eߩ"2Lc z=à-ŐjݡOWu#~"Mli=?r]d.ug^-th %AXaY䢣UecMUq[%SX(^Պ}d_+Ԣ3o43,O9T>́k42[}od&7 Ο?{fSS{.|r:O"?Yy_뙘zCm{ s ]w.ۭO1yuY7/@ ..%B_DhƄCdE7gcw+ u˷X0k*P<"WRlӠ8 x T]ɇd+q*^\w?딕-tlznG4f/^1U%'Й5r Y>jӋ/4a[ Xnoqȣ!J #Ť] o N'CডC9aor A-|Z8k1fGy(BEkIAOBghD'$7, )!JɋmKlœZz4YS:AށQj3nk+؂mYHn YcHQte CW`=W_<GCW:@UGk`:5>Mǥf4nj_ﺺ$69Eū?|ȥdWsHQ2O)MKT*^$lF.4cM̂6Մ &g](sR:-!iǀڻ4 .-Ò6"ShA&bCS2;ʖY "ga X,' SַY4EF>XŃtoƎ;Z׶?t8B㬾! Y8f?{c=Rtvb[(y㌁a| / YZ*J{#7'}6֡>zȣ 4^͑) aH-!GI(lA# ޅ[AD#X 2T,v 7hDXEϘ%t–%3Bq.ٞyѤ>06_Aݼ%;&fj YJWd ^XSˑG9!:Qf6W*vzȣHŜ]m-!7߉Ɗ>Ön5mVkg#`׏< XC=W󖎛Ca}勬aX濗(Yk2z/sjs֏>>ar,QQf[菎EhτEEvt9C-,zCƭ9㷷z53yzlF##%8 F ψRu,0QVsn{=wȣ;egbC3*_}~{wPb8r<0Us[q0a$xa;K 7)7y⣎N8hb `tȣQLlcKwVLt|' Y 2{#=ĸV~1>.q(f'7T;Xje\Nԋ-WccFҵoW jUw&IMKAs'G{\D^ ;^!{;LԾ_YQ&G 㹜MN;yx}jTp{=/k|sˉHc~}9rwopžmWA?rTLKyK-l*].P3ҳ?,Agcٙ1jkzժ!X۠GRŸ́1Ίd {!G 3.zp?_e49Q/V C+BS«''WQ[vvb43d^/a1k*:-E~-cl~qSꝫbCFhΠϠ~Qz#-̡ٯ;e ~Q-#1˯>4m^1^N/nˬUoѵ,zEٯ옖u=8g%|$dzެtzsVG$vl~"zd$9:etǦxm}̻q( A{Kӿ,}xVЫ>9f|O=[yyZn/Ty{$Qslf}jGdGk#ʮ Qr䝂ـ j%i' ,1?SҠ> " HL $ VGQ.SOI &-׽o"xɢ5_rwm= UE|ɋsyn2Z^)M쿇`>ȱ# Mssb-vV#e 6 2a/F|~ }osH7PT"_ Tɏw)~-L-, -F|᷋${Ng$-;F} (3l~cB+:0zO?TW gM?Uq.l!n,%.>`ah Y< FBH52n{&1p Ibymbt)ڒ EQ` Q- L*$//]pRm|j&ѱ:ҐBѾT7&F@r=>f|Wp诗rvUTZqEGͫ*ѫ)K8$lHz6` WqPlQHkm%+LvL.&n7Ad/Y!Oehxn~R!s}>X;y$~=$q0}'EwTKh~*h +m+YŸoAR< C0z3>ղcIʖ%Idp-oΜ'om1go#Æ7 RWoT1$[:+ )U딪M,+4Wգ`(v/xE]\Dxy֛JiupR7Qjb Bا sC.q}›jZ 9X6R}Ӏ)e$_s0Q²,s\6_π1sa޷O͆$V4cfEv{";*H' 洏$K BƙjWFIF;< -Ik$avAS/^olT~D#_ͺ6"K}"R VTtVb:O?hR,3G4 a>Tl(x&/wR\rRɉU#9Ysw=y Rzy WKL==Vuyy<7LsZ%b1kֶĜ~/%2*gW k5b:trƖǒR+ HJm%_`(R321ZcY2bT#FV"į(zޫ.z__TCsp 0ճ -l[kb-Z,$y*v/ݓEO/ۺ<]wydzp;̋YGs+os7Y b_0Nmm;5oX`Zr2~ͩ[@􇢚kB޼zN$BGLjoϮSL"h\8y jE C_1A k:pA  ln35[K |fy~)+ ?0v0f"n8QF!)S#St{~>S JscT|qr){ûOJRw”fyX}Hsr"F,y Rc>J&h1hRTC\몓C"ߕeyk:x?=f fʮ W1&:pyU1` SO K 4"3R̓q0 !g;:5L~ b ԟ5Fh(ͥݨ +&F#U;UB{Z|ϭJhW&8SJ/GKb@㙽A O~S{Hp/޲7qS$SASm~OTl|J|ɔsO\4ښ!-@lpTMc*W@3rkku|X%3qr4|GC 9SI)[fV!}_ȧŖ!#PXˍ/5`*FMmA30QtVlEԽʈOW )5\LΦF=b)ymnsPW fi=YWKLjgY 1qMțmb] Ek`ӌ۶㿮5 E&$P=&'dLi9蝷 &3+[9u2pYJȑ"z޽mۆ RHվDz5ٜ4J:(d]*-&kFN6TC~Fo?b͏ rXpmD2PnJ6XXt!].ì] qTCL]dl9=EׄrM?yY!L1* ޗ`j6}pxݞ g:ǎ|鮌dBkf,vXE{|íJWL}7x= 7MDlAl mE?D^އE;ばkԻ֓Lyw{ANN PUB 9w*9lSf"KuwɻEZNm~ѻ\p63W,>P+{3[XCpu)EGϑB6xU?"nMu 7L'` uS{쀝`5Χ-iɞCdrM98oCݲM˭wKHm ѧm-%芐[$ ;rP?N}B@X={]_ʭFo+.O/>%m3wfބ/Ͽ3_E'fWwFhYoev<E1S  ͨ+<}p*)vտWܞvɰA<~ʩwg3_sշ,*f?`o snW::ƚ1c>Z̈́ɔFRҺ!#l)Pv)Zt PS'뻭k$DƯi;9$r!'~LzJd|D7 Sߋ \5l=}o}?{q^Mƙu%,m=73}F]jg'\639>9›Ibu|ϼ?,o<'bF<2z*lm>7VM7lA#/o l4k 6NP ^~nͽ=|Eh1ؒ4Feyu(,A3ҼyDw;赦8jRBj=UAM8V<ƈr?]rY=Ǎ2tS֔\2c^j*1jB `*:g%gJŸk|>R3q\[[@Kփ ٝXz$j8ϙz y5/zc)tJ œDSR PiK5zh;NYAo>\F/a!3NOgܜРJ ]49h*+HА$b˟Hܢk;0nc4_5ꅁn/t.82mb\׎IP+RfbtŴ3+>/<.炡w YZ Ʊ&IB)VmPXt9JdA&Xѧ 9UWds d:7)vi8) Ln8rq[ `"K) /`!M9ܒ[>ƔC=<]f.Q͟wa  ],7#,^R7-Ǭ:Mk0UƦʌ!L6)d Kf5%Rms kZF#q4 z0{R,Ih>kfçU}Clj[f1y#F絻ɻ5/%`);j=e;Zi/v9a(ބAT ,c`Hr.JH> IJS8e{GpYB1ƛ= *v7T")ޥ]״X{g߰.<sA<τҰ@IAuMț4CBHҊ2%(td噉S͋g_k2W ikه&1f|6hܤ9%]-CC+S l]?4eI3 V:*v* fvU\WaZY[uwɫs൲rE1zS5!{jI px$ИV9ev٬Y2>|%GS:tK|:I^4?amօRpDՀ_o`mF̮4o75)oYoJRChH]b OfHakîE$e5[''d.eYP/ҰGW1?>B΁*H,].jK\SaiMQ;&0*5b+nY_RÜ3hL` *QzGeS%yny{>ًѹhDi٤k.vuQ {o>rwɫ U8%D'i |&F[0;fuj[zWZ >ή%ZotOZ8j<Ҙ@{ Pǎ9Tblg-m-16&֋>Ԉ=\ (w-jKIC:7MQ36cxӎHޒi۠Ai,<Gn4֋>i^n WLs돟~ߤ%g}-WgR-пqb>xtCgyC_˩Ƀ龜n¶ë s7,㤬B$)l1jM0tp=ZUF ]003Thm)!6hVsO続,8A= lM^\4Y$1ezμtiwG<Gف~_PچBJ 6証%GoPkpv7Ӟ*W7`Y-<] |g/iF-W lP=Z[ԄFJTF4h ail4wƭQ5w:| _H\|rv2Rkװ(u뤤9ܞ,kM+?}R?,W/;8 3O~w߭x|JR=_ ]A+zloXҀcXk1U.:4{AưJF$ Rb{ZSb:5^?-q=n` ͙UYu6 B]ڜ[ "[&.:g=Vs&C4(g2[of--NJ 4.\~w癳6oGLa)4 Oם~eǒZdXE3ū'd"uŠ5x rR1^Fm-Wl'SSt? "曅١vOh4zrÞ80Zi883zsM;nܑhVl_׺\ oΒG. v¨'z8Ly+6 5kέ$R#$ 4JM($,jҶJl-7@+jBp%ْ1吥 E%rJ2MNU/߱/9qC{ax-C)ߟvGhHEpiRbM5s9SsF\6jNbu䇐QYݭto^P+GjjvFUGb0y玲y75<=:cTcjWLe㟌w15V7N!(."s (~҂a(5 -nJ`^ 2e+2`IքJh0IBT1 ؤlFFdPf 50$2) ]ր֊Aj*%9C10*ZMU*)gS! ɒ--*`2rF+HBT!I'xo0A[7l^LκXTEJU`QFCGy}=d V )Mҕ `=յu2E^{aLXR͖mhDw.򮂬@4j阌*3J"B5 .%0ywQ=kG}zm4{/ˠ !c[p5NkA:%oX<*Tơ!Wtb'u b̅hDs qAdVs|b9U}P>D}:wl>wyeO5sZlA4`j1ZH.[&A9:=B7KƲM=TZ~IG,MLadV>12k-^ UW^Ekx݆^gBPd̎U) M'e1,РvyEGJ(Ɉuqh2*߲kUܠ<ʱD)4Y Tٵ/|hYk@y4-mʣf|]ϼPpY{!5k+hγ:mx\ B W7+J=y֏cQB YմA8c b6&O3,Fd.dBUq UI8][cK*6@h f_,RpJZ:Vcr*ڲSbE$/qtqiG^x5?g cJ[Y/gv7x$(+X5"gD1 #vH>PނZj[SE>FRb9B@]I<:$SCT2[k&@B˔#>-VV.2dQ,c>v cARU-MF*Os}a7q*@gwv䴖4)*ɥzPc\:=hP[F#|:]'6Z0~[8*WД6s^3SQS"elTRT*0CMd R2[KSM@BS .𳿢qs? ));4Cޣ7*U|BlMDɢ/OlVY,+StИ}拶*XS^W) G7$-kwYn'{D>;~,}&3XcZ>9o%а tO6eo'3U A8V+36|L+7D6hScB 5z*A>Fo.{JڝF['_.XlY^`n-kC/},rNIg"2,вqY3 d6h0 0ٴzdsJFr1(U(DS- d21QBIJ&oF.l0ʞDeS3I:g] p{K85w`2q/zE m=!I·l&>F Lf/.&7'gWO U L Yzk|1UyI-ZGn\2tٹ=8n!#i4Ð.7rX#30qnxw}d&Z,dV+V2FU+WuBH_UZOVѪwS1( FN"#)!^z43\󐯄Ruė:tмH(L:FV> Wab!/P*8"I?ޱw&`i_Z|ԛ@s1-(3a=6sT0Vk*P*KP |z ro,@s6]Qto=kq'_V^uܹp4#+u5]=j[#p~غ[lF- Qd%D6Glwǜ92n5bep"'0VV)PaU S5vh־ e1؆ <Pnpfԣi,M3Ij]R|mWSNrHޭۮٸ ?C{(Y-߷' X /N~'< [s"uU|u׼qWK};2yYt%od=x rrźJx,A=7v1{'w"+WƾPG?͏̓Q6b.a>^2:8jr|1xCl',Kpʮ2i7%\ ??𚦹F>8mnֺw~ξuom }En[Aa*]NG 1/gw,,slO;pd` 1};H}}]?Jiko=bE*L^D^T@ۣ+$7%-G>K]1!dЋўt0zا}뜭|mcL*~\\:#mw 'W"l%0Ɖ5X-%ٺ_nsMm)׻O}m3 dv5j#F@1=3?/:Iӄ_V61G`)Spx5a}撥__k86c?wmY_!eiC?'@f ȀD9$YWݤԼn[dKz{VPuEmj#0Fd`DKHz/XPHĠ?l> ӗM;s_T緳C,m[XV~ʷ9yΒ/-D>=4;;0.$P;r`* [ 8 u"Mz97XC|n "-H{Bb@-ݲD5@U [e}0qjN$=T%vc?11aqPspȩTB]o!A 'RREWz_a0q]J|o7s7ѕB g6gpR~7=U( Vo%;~_'0L"-tV_]QΫFw~>aaF(!gkRXLzMPo5 JN ]4[nh9^R&F1 v0q+)NÚs!gDCL "HZ1- vb7J vb71-8Q9$ohZKKVf8()9-2qI"9'1E0-r_~GK1M JD8ޢv TBPL*bD;A%채|)K _R!_s<ǘQcøP!#UjNNd 03q~ 恰G+ԥa{k E=$Gϑ.'`Y5*Jṛ9l ,M#55jO GXJ"ˈ !Imup38 py!`#O`/(8ѓQ6y|lV/s}( LIF;m;~5S<5Ԉ9MjN!.zMuf L ^_A !5o؈vIymEVVٷISX܅SvHmAm oߏ~{oaOM2? /IDa$8k24]խD_4Rv\G&0Q)Ey`*-s:c @@ TVؚr|?=ݾ*Yt ;dwQl/ &^E[ښ7Zr/]MDҷaM٠sLd ZW.\ n2X,7 yGrTJ%u:L>TOh]ؗ&mS㾯N?C(Mdz:ΧēZ,GjR~Fe Gx:*\0QnUaV1* I,.i-v( & wpr4C3)by$ ."Z;9ҒFb)3H9 G(,֍ZZ´?^9^"Ȭ%ց?^r$iM@;Ȍ▱gΰ *Q@Tȧ5بH&T@$rıFP.yd'~ͅCʡ-v dnҰIY\ȣ S-D# {"s"A(dV%Ji'G|G*uHUaMTP 6`A!QAVm{\+`M9$8YRa6*#򞁠< gpAr *278r0w αV * 0;) hpd,x(\HzH\XEB2̎G–*14ѱbj@c$Re)Vrئ(ad=Tz*I0zAve8ƼrȈ1]AO>+_{kY (d4ѕu4&Wo$yKTSRwzq9:04=ynyYm ~sRoMr"z:/8ثo௸({W+3u!ڛ3^c另n[΄ x  X*-@?sImτ/Y I&k 2ild bf#%>~|fa!!A'C8=4(ZmT> 8o@>x Xx5VP8bPb/)hxCVUwg^|"2bw  ٽU6[вg393+k3=x4syMy"G*y"G*cR~b^F+4r9x* P(k/p gbŠYLJϼTGH{~YUɒMw"UB[YgHA>C3$!b'T[ "OK%j\UWuln?o> 'A&s (`w㿽Gn8=jz#>4zs56~VHYQrYEE5⭞0Dҩ~rBGЎL?1-?.7 dGX&K[RpeN&*YxÑ0;B͆KƦv(G@$x[zfwћƂ2>وv1u^Sg1ۭ>y(Zt1,INЬc~P%5(S<(dtRWra3*m%w&wWLF3:}vкʍV# ׅ7XS,4T%R0Uȁ~cvTC[g8~ /8{({5CPuYg5t0k]Fj@u0k/-|gK9S?>ɘԲ:&TTR|pkf0m?V6JAc i>;?^>5+DM4At<K t1j -kå1;5tk1%[]:u&ScƓ9|.Jۦ38~ዎ+$}W$CZ2ER~Rۭ}fbQv^ZAGBXZ%"D1ʓ[m{5&Oފݭ fq+5Dxs9ͥ`GՏ>B7vpO\Fa(7XtauBC$ь+z+M+%qa8}Z{g`}4T܄@I~a=h\#yJF5?R槦_B} &z't{?HMO0QB}H#R @pVHu9#r @VoJfR1 ϵ LG`,x u󦋣>}K JVvHή.A8--࿥ %4J*hTШ@ޕq$BevF~bޙyYCS&,&)0}#W)fuU!PfUwddD5Hdl͵B gN&+і o?9lu>?9?ʊzѦPjΣ+q9Γ_;齁"moZoqʀ6s!1KFhBB5fCF^k-丟_4_2HZܑy7Çơ՝n ΉhN ^s𚻰kӼf=k ;;bg&o,V(8.WpETJrU ޱEwSISD]&$ 8EI, RdEDllTQȠKfuNu ՠw}F6Ct'#M?a:CהъPz7 \ʗ%F0tGk7I@ȆgkKOc,DV92΁3[ɭ5Y-4@,d:Y,wi^?Sgi~g?suOv98ONei?9s 61Ho uhr\1yњ6N ] f$>u'ػ6{ʧ_^~b׹b'&i>vz7M$~vbr=tni2dhh8z`GQ:>a 4I24pp C;UV8ء>v!&C+}Я|RYgQDTH'(]-q1_U{-rXeu [})-_{u::>m˵땓w7is_q̎D;xfm#!d_iWt #]_g;8 ]\ǗW| =UJ>1KgSe]wʂvK;WSAy‹.lWJfhlS֛ P2H(@TĚ<;/[hkA%*UQ-Nfo*0~>sUdn^{Qc֪~?f4|hL;ɚΰ*;͜ M>k_؉'xhԣFIQG'<M0ۻˏNG[?i7G9i;[nj ߰~kOL5;Y OfP4 %$zϋJ fBl(IFC3ƏІ;oG7{> ˫@Z_<#g >q$0r#_>N%`œp๯Z~߹]{4+{|l?xcW*{3cNg̤Mz7~v*n_L5بSk5ƥG<<~NVn5=5?kZ; K 1]ymmVnio&B17A^RkEmK&& U~7:JUMhp/ߙ%񎿔GrIk_," ̎Rեq:A,PO4l~M2{MOݤ ڴ|#*cO/]\sKG\.>kN5.,lMBx\U.9S rQh KBX4 ,ﬨb(4ߝ%OXIe F\VPB9YXnA>B1p큱is 5%L.(޽761YSna ^$U) Oq{ڥ Z!+U,^*DkښȖ(N̋I2b E *F\!P͟ =j87w==e|&9-sㄷYp6pXgZ2\>k$N,^&Tq*jd)l}J|f"M}sۋ_8). ." O\abE;'t^<,m  |4U5`Rg*1Cb!G-GFv֦1OѨ9?\\Wa1XQ[Vq#E&+pՙVj@(~wκ̮X%>BIm0S œPLbS$r``zƑ=QC{85@S?6{L=@ gᡮx-C|ihp <M 5) ]d7e"qC{ :}ME8Pe.\ɡ^"d\_KrH8n.A7[ ngzUĪWzWt٥ee :JxoXcgjXl:{XJ> UK R M3Q5_+kL0àTm+Mc6/[|i h$h2SK@+E+" b¬ݔ D#My-TniU'pg馯s1ty2D-2 QCc>HYKPQ('q}C삋հ6T.Pռ靲vfc닠<ĭU'zK0iH"AHۀn;OQk-(K#iU5KV$[h)ucmѦKIq cZ!m+%(bP/+={мvѓk]Tט _r)^anS?8==:NոmZj#|pr-ڌ#qo#&WZh*b8AUBYb>Ԯ7ƺQ1CvFGDp.D>6?Ok~cb_֚߿܊nhO99k]6aI};ۉ 6M~(}ʛrrZF ؾo?pyyDeֿI 4'1mf2 Jv#| }@'V΃mYCgeՌ[y)zAM>Cy?Ⱥ-ȟ[o[gKY1$zqv"At먱xt_}ltՑ2,YKxv/%_yV߸8;ZLWV0_o]9˸Az9TŤJYFm͹%l \Eb(T68FnT{Q扶.mmrq\󟯮;e Ym%ԳnYo KOV}^=O:Wr$ݔ٘Rփy7!Rn  m sC#Tz=%=;y$)XTf^Z[GFEP:~2҉V$x\Lw;}j~Ňzv<۽@1橍dl;f=\ +6*$m@[0nLmzZ? x!^owQ8xh"cޡz46_?ܮ?+qhfh4}ͮWw);jG gq]XRjsԟZ`ߡ]_|w8|;лZebk Ow6˦jwe>XGBut?*GFNJf`,c6Hd2{lSiqU ~|_%v丄Ɏ-]TkdFW؄v[lCV\#Qu&54:N[O<" (OSּ>D ?M l?pFּg6Gpƚz8$0zߣY~m2mܛ_zܙ0 ^.1I.q?6ߪݭA{ws](a-ߒ&\rE\` רɘgF> s^$pz^**BsjҶhlII9:H~ԴTj24-fzxLtK2) @r78L N7JĤ$[u4w\aH~Nktq)&K!_nb`1)οOQTK,.E#-. 8v 9L dJ&Ӂ d$&^7_ɿM L@t,NjliRN[6gAˇPHdpbGD: 5 KO.eZ ,°L~w #r`m gP-B$HF=˲F{I{c4̙łMebE|⩳j!iJ@9C[^K*$CkR w@LaYЎ-sOBP9&+k BH5=j V Lvr8 ״,!t3T*80=͞aI՘KԄt|ԈIDxIMl)ܐ/izH5bJĤ䎑4# 2m.iVCC޸fl`r1Q1Rی-SojhuCC޸K/Y"231J#)|=0T4e4:| y*S ݲbS Hҁ6RMp# q ke^ZΰdSB1,ݜ}rGW'*ju0bi,L0:&8ͅà#>u#OE` 0F6gPM1J"׿j 7Aq#e+(#M["׿14䍫hNQL[7ى&Ѻb:cX+/fݲnchW,duq2AT1jθ ;f4UtNi%,)| Zʠ0aM\Tlpoc^G2"I!C,xC,pS`1H73X+kƜ)X-,xCLD2`"`bJ-i^+ň&|%p$(i^jpByQ]# :Z l`I*Wd%zx,NiǏJ#A.XRՉ(OTX"/quS37 ]vt*HKaщzRp]`tq)&cJ %DFt/84õ#;} TA_άeQnOT"E~ +0aL@iOI̗'ce۩rs]=7R֬Z k*AuU{'*gaG'&o91aU],prBƎ cA jF'&_3K?/@S- DvzlgLAf׵T\b@9"ş4lȁ^٧5ք D"B<1(!y$Gb'X,RJ@I@GTH5BMH@QD=jly@NX0 kV׮1'TU<{a6i"xĒ&9Lĥp0KVve#~671+zu㚘/wk?>I{46_Y-vozޮ&s,y +@lWKr3P]tVeE;mх-&0MM1{ Dv^['03梥B +UE`7Ԣe0-7 գcѐˢK`3Ԓ=~*Qn &MhpR=|`ʙh)hZr&l_gS޺n&ĐK%j˸D-^tk~[0?ԂeНnm{=}}C8>IZ7*rBx+Cj/WԜ'uީX1?yȹ%C,VH._GmS5 =?鱙gswģd#ŘG { ؒ\/Uҧ΃D}$EŅMTx{Jp ;! >]\ʧ+'IRꅯ'oS/XZ~Z12|{?o k% CY%L=K!%khvcۯ,c#he]]dsނ’fMW~LNvI'Ԩ8)Z_4qJp.8)TXMՌ}E6jIbl ZJܤq=r8GXۤhImIRrbiS!1f ,:)FMh:)Lo͒4:X01>y$KAIʒ=rOkeHay+ѐ-6ˇq[aCq]dkfsM| 9t >ϭ!zOh_>gcX:[UBϡ`+,wNɣ= E$Y8!Ni˹רD9LרDvp^|Ÿ~tM*)=g5p A:S(ř8 !j1k$04I,Q g ELfH4YciX9uuYP$z^Y8!g%cz^9fɘ\νct+gǏ1K1\x]9M8^x)seL%2& S[֕ T2$n(ٝzx~;%_f[#ΛϿ ) -9[΍JvP&Hr{O&8҄Sn+ ߆oMP77\8-ڮsft/9ژh/ p L ϟP9J5W⢀0类k>V'ks8t*>_^9^hÂT mG7K_2Φ*5ds8Ђϱ~aز/FuaePNd^a[0g8M]=8cN׫ Â\J]bd?*n+mYMSt40KX8b҇~yiZvJnjӖmmLVz>W\V~8S TQ6@3D%z諓"33%|$z(61EpcNBՉXYEQfw{e%Y;2Ph$ꤩDfRs=^&'L~Tg:N|ut Հ27NQÒ9<&ޮïNT.a̜s_@^weqHy a`yYvոc<]8fewÊވSDFd<vܘ. wezTh+>ae WݟW /581.Axa}{w7)۳A{S߆ 7_$bqSJjC01嫟LvK`0F7j2Нd)`bj \wy{,1G}oQZ)tM~\k5¼!RP-{t)76_.Ƀ>WUNs?:yvrac%'e/BPF$s2ExpH9ARn3N_ݬR g )MAP2`8wz|A K0 `VU϶i'* iyvy 5,θ{947P>*USm&K'P,7V-±g\xt;~J3qJsY~07jb6}K!1[, ^hǀ)So CcJJݑOk3`㧓xm (Q-w軕ۀ5 cKxQR8,(ShaX=@c(97 $S7 hDZJP9EVD;nE2ΐ1 8B̉ rJc\9xm^q! D!< i4q wrļ vZEK݈}i/2/R Ԃf/ͷ106w_\dfr_2/ $K=<_ ___zFIR)ȫujOv# "L7>tXn澴*N*?3̥[}e>s|cdGۧ'Wg?}*a `x5{ QXy?U,we$T7cկ\ۏaqs"qQkV dӃO2`,wja\8VgtP00E'\@OVH_hYxS(s,ZD¼!Vc0$3 ڂ5]2pGbr}uu?-T .ܕ#&ۣFL0Q9qSPŧ,X$` 5nU(%&X`)mdHE{gtX,<72 D I7bR}2g\&Kbީ$ug+<ĤpdjSVX66ipS8 ARVPZ¬0WqQx`^ ;v$ǝ hzI r]f(CBOU0Vac ՌR8's; & ڀol$0sVÌ$: V1_#/y؁ A PEk0;Ay"d Qə`b^J s} C뽗 qQ\8 }"KV`#ńK#s[2@Cn)a*-Q/09PzKF m)Uݖџp*,ng.u}߮擉a~46J_ wT0FIWfiyG"gDѶM:rW <:*HB~h{tTWV1.֏Բ;:uxyO$GޓƝ!z~FGWsxyO4CE8znHSaϯ2ĝY닸urO[W7 E飷盔+MZqo`~*`߅HSVXC i3!g;F#)+]%; isa~*Ӣd[2Kv9Pǘ l|FJ(NDaHT꨼wJC4GYxMԱ<U7x& t?5 OI6=PwT!O* 9cR Izn朌 $Q* ٝuOOG*Ԩl$y͗O~{(tl՛>J;ַb+8"jŨjͷ8^. %CvYw30*R)ȂRr  `kưG_V)L({ wtG|X`'Wڗ+qm<|,>GQ  {:uwT<𩿏S. wu CPJ h c*d! S]g˽N?fq7]T'i7:SbJ nrg o’=7 mRF_ZB:i'' d+&v dbR ͌m;@vOZWh~ÀyL PVa.A΅ڹ’U0t *QF;!S+FO,8 Ob jʤ $%UwJJ"\0 $(j,u+f2yvZ6T!N=IvFqtOl0fAw!0E6l R \qlD>N`_y/XPK ʄP~[#+[(5S/G)QqW+w~́SWּ 8.&2x%L O CcLXtG@aPHZ^hD<N!%S/[pidgnCH9dywl`Fk"2d=a3Qq("`2F{^HPYI֥ߗ-yia_qʋ-WܔNgl)ZoE^?X%mvW7g'1EdZ݄#h4'&o2S |OL>si_gWϡ|lྶEr8b0}zru#!TakN&c@u;0y'J2 `eC` ,>7f5[-GLR8g)DV'Ġk4Y9YĘt g2ə?_g_=j(ʊk^Օݤ2nO*G=;qh09qpow}'E4|nw6aqA/uv\*?ȭcءu+?{n1&=NtQJ<E0/o0DfJ̓R(EC",*T. [g.Y2U9U8x#Y嶓[.)[G6G䔱޴[vGS[E4KmM34햋AQG IZew4U!!\DdJ4ː!y\v8a$#Xb}E 1!lAQ¹R.0EƸ.7sbCgRѝƯg,XPey&4]AKEG1(YUaX&j@P)(H*V# [dPufm("KEa2c0\²$ܴ7YOl' ˭O2%Vn '蓴!!\DsdJ6Iz>ixc[.)[G6G*ԛvhvkCBfɔk7VZE)$iABfɔ`Ik7O )XYGxpp xr E4K*VX-·ڍ GqsD.)[G6AhvkCB6).&;v~{/iQڡ"=ib594O)kT_PwemI kﰁjcQ=fX#@H߬H6Al $XU_VV]MtFn?i򘧪QcxŦYwK)#e tnTir0Lq4I/ɅF˨Pr JATm659% mGy9kmgit&XLBޣ{͒T?Z;cX29u.R/l:Gog0dI>KYO~-ʕSZsDH]f9ȧ{atfg`9AͿLFyRA~$.S*jGiMx3QeH;ب︝: A|G`-1#^Ɍ`2IJ*%s|[GKX,i*7C.UJu.g@ ήǀ Cj@@1D7),`H{EOl6HϭSy}~oU sh&篎:Wf3t}u d>H| &UFhxpWt"M' _G~*#,#^V$^a4]jj93a6 <4dMxhm% Tf OfECƗ$?&p(7h{e) gu-L0B.nA)G-,0}M]>hv`}NLw1J\N[U @ b+`;M /9͏ߖorļ5(XtQ.$R%L 4jD@$':2ý[aXqnxL^|xj CVC3&+8P6՗Y)*;tsq|=rmn~C^,e#l=x HQ+CW Bh WatYZ'd͡Kix2 ɺSz }7q[NZq!KnHaE1C2fE>70j=X.qdPVQr ;*]P%h^ K/OV %s?XIw-zRYBvӗ( ezj%dv%r5"n?5;}H+ ^RH%Ķr|,/WQ*UzhrMy"$lOfgѻ/oҗPV!ų$Ò ~$I.O\r6,xGW!f(H$ &sNzʰXJӀQ"X량w]z$5 j՝ۚKOݭ6?gu&aVRkO^]:ޕ1k.A*JP.uW"Lo+"šeA"6 GL6^8ıf V9-6p@e%<Kmn#}чs*JaR[ˍ RuMrPn(Ry f n_I 2ϴ:Ffډ@o4u2pU;9Z0IJ<2Cp$^.}RjDKoDpXl(y@:*Tг/d=~8FVCB2m92Tmkq4?:u@T $jWs~zgby$4IiӲDc8:d(7TZ01f,6@ј3j1GrDbu'+UeژPxx3c\: {fj"<=Ow,BK~>J'+>7:1$ZD5QcFPL\(#D0@c"XPΫQn63kHE6s+?2b#10l2z| YtF X'8>(J9b)H!V)?ZP%=/Z2?u:<|>5*iJŤBiobjy"Ox1 UO\ioț?b!^B?+z@Ddd &A@6CKi `4 wۭW TVό'?4Ág0fhMRcg`_/})Q`#kLI׬]0KIѓk[G0tTIYӿ qQX{u Mј=RhMW 6P|NYf^eJ+2VJF¥x2֕3Ufx@^Ttn8*q9,rFQ ?VpVh.j]K{\xPC^ K<KSGՊP;ᗬ(ޅbg/Q?)&K.pEpS۶zL?(AqL n=a$%A0k2o)nqboߋ"<$mVNP3"lc`11Dc S^YmYʹeȲ9+@P9̬!ER koF(]cܦ&E5#CUy +<GK&2€zH J2L e9]1ĻGŨ.7kbl/yx.`B 7AX,hy؟EGyJbN5 KlDD5Ht1?ֹc6KDg~Y+;+Ͽk/q67V*m%`0wHsIM<Y Ok;u[2o@)S@Fa"6MCt{K5C}%A;tQM$ Vd@l`bE/` /8QȊC-S( ~FƂj5}MjJmLk"멽KM3;5t)rm"!.wAKcTַ)9 isqYtQHyFI~No8T(Y`ȸЂiǣ?隻s7ʳAAZGL#:raBS$S0"1JTE}'y0msx1-R`)[A4'c~n 00,\fϞ ˊz1!p`7S̼p6p)U"2->eEbQR TψsbG G9kH`*8c1 as1Ҝ'GRDp1Z{efb(, "Qk#F(HT6jŢJ8D3Y=!c^*=m"YHɕ0>qC3-}OT.dZAϟTEe`["+(P48}Ix?seFYop#$KM;Q  [8xMv3[p' ~S4\*!|܇JC>[/ 4|c$;"ؘO@.D]GR 3l Ss_w !v'l/z3bM](=KiO¬T7sj32qFq/طGΫ`1~> N'U7 [4'mE۠Cъ(&7߀TbTeZ?VI_~)&eY"YFB"pip[07-U~fm*ʺ/F—}=_3p1e$'RlXؽǿ6KD CϟNꎇoK1G6ٯ_a3L?9,t]"5,kK9m[:Gv WfW2,psC:U yigAG$F44AVZ)H{7﷨D7/y.FTLvmq'hRb޲95*Dm׃?wg5. [xCp{AڧR.4m"(azidIvqPхO;]aA?؁]@(_kFܼOt&`tAtAnh ".|e|%-fQ<Y*>6*{),&2_+Q"C%x(q~,LoBYzٽ'2>=?FT?@GR @0k" ]4ՙA^dh-L;3w'j(X)5#ǘw YA \0v+5L) @ĕXhnkyy@I9ezz/ΞĴfc Ї CBږ^UBbnɶc{`1Z J݄cAxYND El&,kC\V.Ųlʭ WWBpY ~mF78+|sqٵ54qr֬:\Q^LN蓼NyAa:D Tw* ,JH+H(>^D#b+ׯ0l-"_mUnn'uȗ*c:(M׷E)xQ=u ͋XJ9a !Fn1)X fպ:lQ$zPZL4"S QŘx'aۃ tjAc~(:~Vpp?e!CQTa jn&``m@bUm{V rx%hMAWTN0E(?k5xWC#oJ쫃 ^&Q"RNkG\ 3[(A kmI0;pA n1,p hXYRRr I=h=,$%y )jibcBj+ 2E`r`4uRFk*Y"vjtNR'%tXIO_u/}ZXՆ2G%$Dpw0q豛ќqe( B@2# -a 3DZaLQdU9b$By^$ :vy}%q!q^_E a>^ ؛Z(qDxky^l= = aH7ԉ!$iӱp"!L0" Ё#d0;+=kx_j#w]R7 7d$CekGT=hvT='Jn GV;uggfN xE)stR|iʡRkzXwMV1XcW@e{i&g3WyXq4~I-L՜^BEcWHn:$ 7ڭZGhgيrseYWz. յ8Kijr*?B]Zkz$_=HK[aШKSEQ -J"zƿ*VRQ$D&1%(6$YP~,8(IeyDF;UvҞj VL63YE0D+IEob \jĹ *dOOPTT"̼8#?!INQɔw㲥Tee+Dw'uي4VRһFR{]\s ͓|Dq> ‘(<)!Ux.VQ#F !ڌ*ib\%φ֌@ I0N'7nf~qc-l BAxxm wnvZa鞓8#LqhWْbWэ5~p|{; fwC8'= >+-E\^jD fym*B6؆!K.B[5XHDI%Ե)wJ k_;]UIB]ʫ?@Sf:{w6^^Lq[T)|zMDAEA$|=6\ tlrl0fj\8Βm"}Vve&mf/v<]a4=RQTa^!#U)=R{CJS!D*%ʍC$ `o>>'fv2n_G8/b -ٴ| ,D hi3Zjw%&zeZLp;HžW ]q4(CҔb{;}sq-.b]]?F0.JJ5YI:vf,6{~}d:pBáAL(#(Dm̸Y|9J㮍وsBbmPDs$Q"pd֊0M%wA'Q/oP!2>(I;*͍u#ZaZja4RP3 8*p؄S)=*EUhdSO5ON~>tnyd~;74OU gDD.)߫&&& }:'3O%Z6z4{7ҹ(jf%f8߳Oݐ^ g; _dvX\g㸂1og|5'S1D .Z߻YvhqG4Z\aVi4hU@KhIR瑛)Z`A`UPB<,qt{ 5ax>)E`ҧ$ȥdU9eJD?|R\*n>x)EG~QrQ5^ѺX흉e\BH{dx>y~'[Jco+AW߾"' gCA 3΁NkAdt#! y:*%Ł{x*=dغ$_yQ%d)_#~n9e^I|IWpmrYVYfvQn!$3ӡ+U7{m͜]^1_^ 73oL5 ! (9 p7(ݳG,5_?,_y92~q޲~MljG߃QܿǝUdu#_nu~~Ma ӬѠ4?06tbު 4~ts}0S4> v* \ &g>ܹͲ\I 롮IgMo (X#CFPEҸzx'&2WƠPRDJ"a.PnQ4gEnzˬ?m?H*w(Zgx > σtUap;K,a'TW"T3 _M#yvq'+Tqzkr17˝gsr!QBC"৤XAYL'H&<(!Z0 I8"I""D<2,кO.rN:9DvTu<>ш}оAV%0,̘sX95[D(\'ƢHd)6.P"]g%*"`E!&:ps+PTX.hKԃ?P6C?!b- #$s U"c#UL2֝Bku_CRz@'qB`0X: Ǽk$@ގUVf!oèkj=jd6w|n%5)ӥ‘QDj_Q/y_uWm?I:F&d4*It(cC"I̹BE8$LaWDD!R"ו+}(&a) Ăe6cfQ± %HSQF4j'ӆi$2B:IhZl&$j J1ՄO>q5ǡ40t0M$<&FpeiH ڦI=_-|o$T'g0XsfbtPdR*$Q:)p5TF)-$%DMꉏG1X0 P'&X[e*nVkgޢJw$8'ޢ|;B\M^rKF2N*5vi6EVUn}] $c_o T. 9qq?fmr D}ȷIkCXO}+X_fס.1ۏ:&fbø!89u>㻫3TXF#7tu67 qR gV2u˯.T;L>TS mj==\5 fв80v 70#h3\σG'nm[+>5@%@A'hھ,^gmkٻ|GJPdSFSI+f8 3rEC@Foorš͔ftԏsJT sA kOBA`5g&cU1&a(d DC&!ILhx9xᑴ$h e"$ c)%bC< %QD)C(cVʤ >7z֬"uʹx=rR*4IWu*_w+ ;c:Ŕ0 Q#Hip=X 9e? y䢓)E`6k=h(jYߌRs`l^7bϔ uؤAcIv$g-j9`5%XYsЍAa `3>fݼ 6iЦ܆*k$NWBۜOt#0.\m1nz7N3<>e|,LDّ+-I4 Dʹ^-TUnMh u)uh_(aIITDanu<U0LDQq5w`0KU=PlI[5'<>;*6f\&_/{D1A5ψִs:B=LLun:2Bk%M NXto֤J"s e-aT wwTs(me>j62yuR*X>cڧ1.O]ܧ$qIt2J41̲tyX׵5wTn8ܹhu34Ms=d"t?Ag}WB4ӪiB@ԁ?eoRED'R{$}]k4'ȭ\=Ƿ2EuWfQeG-eLi=ErAMvr#?, vK0ib%QVS$wb6p"(7!CcFUԞ  &6ד=-tN\6H)uVKm^G,I BsdICu+2ENQaYEO#:oNr(M<{frT(dIϑyjߕ\@d~WnfjEe/}x6\mC:w"eWPI\uq"}s>-.S>w{ِ3=v{NKA͚#Z^&SM6_~X,Lu{j@+q;=)o M^4.L{2|ԧ0/>uw-` @XEս+հ=JnKlulNt@5ZMp:L'٪ rIKgl_~՝iMuyBUk5q35VϧY2){..}QrUf_iFoխ 2:|7aVx|b~>k_p7N4KkNjd#G˯>buTEmǯFbNbL푩W^#3\0s;/שIqk7ˠĹ?py?I:Vfbˎu\ާEð_6ܴ<4=zƻH:ǘ+h3( _,`/ OBOI>3Fp ҁ~2o#Ǒm0*>%@ D( A#;H%r/D@@ S `)HfmF_}]P5Q6uUH$LfP}y37 K|'@g7Q:S)U$U=I Α4v{2T*pׇA&&t!q7ciwh6]LtawIߏ]%!N:%}]GD So!.oϾE]Œ||eUӆU>"}L0Uh׋|xks wz4n¹8 lyH4$Bֱ"۞j&O&r lNu*'i{ a"T{CDG Ua R'I8N߫; q"U"(x9k9ox\l@eኒѦᡵf1_EY#|@DzGvԼ-i@x/F6Tk.jc)`9 5 P2F/bҞH|J".?RF;WDW )k!: V.b!`& (8c$~Xq^<&>}G1sZes#Qt5&4m>_ ~R_&onn6ۆ7zQrsaSf@]]b =ހ؜e{= !WQ\MkrwG5 5\un`K>+vC QB}ǧő),D@@:x E7O#\Չ=i Q4nO!8 2ZIl0Ǫ:g^->ZF}pМkR[MN" cFIE^xB^"ܪy]3b0E~qTٍiẩwM7S$e".4 6A=)DC>/QĐW[S{*9X'*'z2MkJs[ܜG ʛff(̀ ̍=')Ss$=8PBDGjln o^*bעlS#;@K#oWY~ S K' fBF'Z %x>4 G-K6HV?9^=7<{\@яF99%1?j͒_auKnV՚_@'ˢk9I,;&]b.]=Y[w.[,Q`XEn*۪\z /Daq{r d#`th!nsNMC'@GaM)k8UzJsa@pͺeV9yB #9ӿLr/kw<iEl2ϘW[zolr,idP`dPKT|_ȝfD_RE}#Bمnrw̲x@R?^칥j0t]b'ɥ==r[uqb,RvshО"4jfOiRZ S)Aܑ7 S, EuFr'Co['k|6:2ϙv#ᙧ~ 4K2G&$GNVBU.=h%y,ddbDsp3T^tcvT% tF,fUJX>!# أFPGUm)^2i22bЉ=dL µ9|M= DB}eK/5L &YMX9)TNG1G16Hf`L(>ٞH@$>'eځ %.BGI̓pȹ%r~ sR>DNQ.Eұ')NNKj4P+嶐\zN^Ǎ!` : m[) SirIԘ9R~ΝG/0ZUi[*WSړ%hxAh-Y*pQ΅[jGNDrʀ !#y$phDHȇH-3#0db c0(ARXIBp 9¨ v%Zת}S|Ĉق a{׎YއƩqEbCScpa* FޜeGY560 bǻĊk X"sz8;8pWTrU1+VTA+r|6|hp)s8̜ D"ts`UXV_cTMFI48]*$y]0|WuQߕaJQ{o:|T<}ap4݇.]3ppCg)]!jV:僠{^#pΘ8A\O7o$d-2tOL^'_gށ.,2a0Qp(zKaɈa5O)اfbO2:d6NvhfG&2.ѯ矚`D޵5m$˩]PsJ)ImmvI%jq,Ik"ŖDof{05jp*7Q4K;ʺ uZ4++C횎&\ScG"ͤPsq@9)Oac,FS<ҙ@k1HCW}ǛUi-c>+<Lff54Y-|/S8]_]af׽?LO0c߻8{&/̳j෋򝿾.'AgL%O ]0suɕϳ/1YO1tLRäx:!3H[խ_j*%]+b ?v)ʯ4b5>A XϨ0]lVۺO o|?8FÑ]$c M*@#WhO;jLOmuB]QU]rdR/(.gϾ (_Kʗ6KՠabDndiu3i[xg6ny#^N{|=%h5A1LL'KsLD 7w.Hνzprf%Zl@0`,C 2 ZNGVT܃F(l_A4C>uPQP9.@Is`Sfq xׇZ>>=fwV !krL5*/E4(gX+M =kk)t<1jP,ZH晢CM_/WJ(M2׹T+:ߍ\-чCa3l%dI^zIrj%![zaWY115z§=vz\* ku> V"GY|s)ȿ\Hyf52H-E M^"L99['R@qKHF LC |TA!,bjPE˻ߵ"9A J6L Hwq@7r@F\&:&61cJd.r1m0Y #r*cgUu*faz VD[g :]8>4[-z}*_W -TDrR n>QԒN;<#2oi,RK%%t|Z6Bw̝Q+1淃^jF?ʻR`nyƷ-jXmIϩ20w:k(jg Ҳu͛et9%xYwqVPqN ܈")& FN?J͕(#:Pj*(#q4^Wj^P p) 1{`>l&vW⯻B|)Zp>`TaA>ݤݠtvnA! /##FhJrcjj:#=NDGCUMIꖨDzut՛>2V16 |=e:q#:W9,8?LsJq#Ib^.e BSSs;jFyfkߕl3dc6颗}}6ŌS+DNw aEͥ%q$ZkVc FHV,3FC((1Zֳ+FDZןԳ+ծ̸6bِ6y3㫄a񢭪tTD0xyF1v ݗȤ)C[zkӦŒtvu?MV2~|k.I_z y食wľ{ las" A!-C/S+Ќ'ϚܔE?IMѨ` -~wwgX38b-}Ϲ7+`\)@WDgH>RR8 2 θ[8̛pp _?-W!?ȇ#pP"q$p<ϳ4Q?B0vq;<^]^b"W m0m3Mco> iٓ=>/5,0_XwJ.WE?4gM-&{  J3mɘ*ز n6Q83a߇p2%od.M |EJabꀯjdEy6‚eTSfo)l(c7 nq 'x'$gB0=7-Lż"uluk4FrUXV`f'eڡ>mmiza>8\G Ҝry+N+G@Kbih5oEnGq/!:Jӈ*$Ftߨk+KTF#BBuc:̂:^GON[Bʑ~gnIFj*kۡ倿Uy} ժq?6]Tx1)NvwGUE-؁}rޏբh/AI-@Tlk}wUt[)=6GfgU&bѾ Pu>w@ PyG)_?{+EQ_+(k6ZUiGV|ƄMI,2nZUdq50Gg)@#^[Mx&/#djXl&eT$cF DV\6̃p6皕F;N[G2g_o%p3NI=W)~}JVV4tL91bꮴQ8/]iu^C En ͻC4~N6 $>W9j68>y[hyds| 3}}&8B2ud$Ӣe sT\;O]ƌSѹwh% o/1Zh\U`Dc1]ohNtA)T:UU(%CUG'jנmw,Hk#xІ+&bʥPogWJkc@KtS%8JSZz#eW*K5NQ Q_+G}]fZW5[kIS{_Rӱ MNYn Mm M 6be p fa0bK\D$XUL#q>Dr0|FiL.P2߷r8G_(T٫˒ÌL^(eJD;.drL׍H)ϯq5$5!l6+ÄY icA,H!2__%4ʠ%шl3,$e1?$zZК|oBo3N=KO.KVy ;Y䑲AA!+9ƃ}&}a߅W{| !,צ$n)pܰC[IbFͽ:\M7M?w).dKyaU:ly j<9GⶎVb%9@9'́u9۫ GOgMC`̚h~f_0M?BvOڂ]\ Ħy4Bowꂂ5--w@,={2)yg^\2"LcnY傢zgEI4Xdg`Acb aa!J5Zh\3&8 I %>\Rk}Z[# GSvrkLg)rEd9! aȉs֐Tp*VNKT`EDZCO"DB) n$J{mxeB@iK`JFXQ (CsF1˱'RCZZ$5 OKXJahNKF&0SQa3$3]|Cs8AYԑs,ua5j95[t0zka}G%|~zayI5гUZ1uxgWxkիwww86Wo>MVmrho'/i( DA"Pdnѳj:G!yE)6Sc[Q55{;cGzhG]pSO86g峥qԦtnHZ}ǰ9gޛ^^eb%V*bfU0jh7 0D=ևG"|>uQ{.z?>SڗW#@(ɿg{ C{FSQÈ4?4ǐY=|bs3cnZtYI|.fZT[Ca19D~w*@t >V#&x13_fqQ.ǢҵR^W?? YzYUmI?tIVҳ~h'\$!.2͓F6V|/IZc=T:_8Xv;mx7~>r6?."mQh ǿX1s1I!883C7bBZG^*d(RAiVJCrR As%2Tda B[ M;k;4ˉЂHE(,s$%*9 RK@!MVyV9񊻄/ /z1:)V T%:0710  9cWwJȏ# #Fʹ, %B! hqi EJP֔pA.KWi~1|,B>Й6$`%3*[+)B%H(0)K3pˆPN ҫhӁC>?.څaqqi~9PeV@/rJ,92@iN!܌~t\+UUP8. kV s@2GbJD]ЀM!%TAZ oH3N]!*ZalFZB AyJ ]% dRbLG uQ,T{JGgN7y?v3:C>za®׳k:S !L-J'J+!+ֳXȑԺ[w.v9~=79T+ɧs67m_ ư`~hxK .L/s-o1zﵑ|[t=H :uH,nB_MDdQEqbUl!x.S8@"%o x!Qև<cYMș;++e̝媉2Ti<׋{78$L\8J`QI∱^T ;bĹЁjH/Nm^,Key#ĪڗGwbk@k:63O-LZЇ}[)[`n:d<\F|*gBtF7כ b]G ilx *!/MK2X7/Fzo G'j]34s΂@\ S7|iW0A嚭FZīM#<߬ +f;7-e0ݣDfI^ZP|ˀ\m6|e*R{ .[h,p>G!$Ka2,PXo>k H(B ^C,DPkc|,͗ /0瓻Pt1OGpz 7 nm_| xsEA8XWOq}l%0@x̱xA@66Oկ nȢOs5/>=Af*p>ZNaѺB=OB )l~fJ(4!3vni9y#}\vKEjgɭC%a5k8dYQBL-d[M:;aQvuNeuzZ>7GDxwB}d>Em=ݷ%jӠs6=ǡ>õ6|jG>n‿({vNA3I/{/I_Ukr^.[Sc0'>vΙ> su]8G;дsd]'0i9EΒnuQȓ9C&uvC(w$$ ` H7bUbn H,ADPP7OBq.Ե.]=2  REFbg`: ďi2Z;(⪃B?kp` GGAa қIqTqPX' #|p0@ٕVC[>}8g:u!yu{8E \2*<0tí;0NѹA//$z~22M谧#鈔 6kRk6)lYMYWd 2oXE[;»8 )VT%:07FTHFL 5`-P$g0Ѥ,'B "Y e.1fyfq0HE.FA4́!ge{J:ص:ԙf V2TZillb!˒`Ɂ2 H(0)K3zPN ҂bQ&$9R1j {H,YF@f#͕ F4( )ĜCldCn{N]LOj17?iI':[dΆ_. DsÏl߹s6 T\-K'$x?@Vt˻*-~Xn 7#XT2&+Cg2?sUQ=_T\Vv-{}N&lIX!.Cͼ,Y E S)5 fd0Ș֒Ŝ{ܝVZt=.)ywC#sw"Ĕy ǹmKN&_-83E!NpA2^΀Ɩ,_׋G.>y;m GA O^sdD*&łЅ!򭖒x;ŭ4&Y]GoG.1BWVg5}-FwEU*/Um|- ^}w=Zµȫ >^L|+dwu/EsdYŊn*x5 Q'.eZm_a˙S9bPKthݢU8j7׉n$EEصݘeqFn>hxY*N)L9Ov!!.N2A3k7i%Z\1#:}4n5" vsh]CB.\Ddg5F:G/kũ1vsh]CB.\DdJg nƼvsŠD\q*`gyu pu)X3k7iesi9bPkct죑-K|ũU::ڶ pu)N{_y`OA({q2uŠ֕c I`h*NW\:Tۺ*V(nD~x0weG3B3R@:*8NW]B.\Ddm4o!ڍ^h7W ꈎkM[gSҝ<Ѻ͇\:ɔǭoFUHq)J@ǵcbS ו@/rOx/c%0V [b+}sPCnF;<'8ÜZRbR̖ו ۾C O=)J P쏚bAR,ו@},)n x] 3&ǜ|^WBciu%P |g+s`Rr2%ߕ`v)(J ǜ|>W pc3cfcN>f+.53%s1{] yo֒"|̼uۘ>f:EM>1b92's1X g-O|>f[GeG1s1'ו 1 H1'וYP9|̂|u%pAccN>f+A_ ,JJ@lpǜ|^WiwlO>cx1}Rv?޵5q+2ŗ_T:NaR*1CV64hȡDJEk YuG%OlbA$o3̆8OakIifA `N &XAWI^W{hyE p2o#t]Ie@7޿=U&"݌-Mslϙ(>S,Gy͌A K :N`dcAJ) /0HVx|8FMXman L{*p5po eטn6f^_t޿n”ꆜCώҤR8>e$ 22%m0PBM"T %mBJl4J>YwPq wd0b< Mڿc&#BK-ჱ)9xf$+3:(M!NѴ8=Q!(Ql#ԪsLHJwM/+([lQAp/]8$R=Øۡ:nd^-a/Fky_w}Oi5 43>&է`:Ai!ə*r$RP-h˟ֽO0ȴq3.)f HM/tMGp<(OvtF=e@ `"7Z`E^h4.[K33t<3{xYxத2۟)UŀQqN96X@4'28 <5RT %*̵X*w>[1i7QIP &:xHxԈN%X"-9J€A jâΚØxT@Rj0BhBN!a*t`p"1 );#8蓣,d/@fBXMCZQ(̺yj`g0'ِB[繆ôX!H +#D4sCt$Fh `BvPvm4yR `F(hU@jcZ)SKp cr#0#A.Od` BXid{8DYpڨT!XX!XيDgdj,FgIN38q ,[WO;z)y tt3EblܗBJ:y:oؤpz(Uׇsʡ1J??%'ן* fٌ7/]h ""G$0XO?_޴O|?|oNg~g8gڧ p FYvJPje_SF]g^g*Ӊ#kʏlk)(ްj'ꌒS9UE'9j$${Fbkty-^Ag0 4 e4l;b֧V)F7~jV)if7W.ER ]G¯0ƌI2բy7l;-ppw*2!mxrCnLGz{f5kj[ι4<NhH;&j$\9m^2HXv 4&)&!ӊ W ]Zɺ,e)qM :b{z˂w鈄Z9mi#uDh9EZ /LX9jզ(L,)b:E炬(,Y c4V\qsaz3;)Wy,-KyYV N}`UOPCG1w neZ(eldB85}I1Es[RɄ4)`Z]+p,[\'d Z-=TԘnB)cjVWz^[*ݍ JJy@\1>XҚq d0߂׏>RZ70׉1ݢRdA?Ӑ圠GcJ[ Pf[;O_6ි}˳ҚyYB`K% LA_dTpYFUFs43k‹]yw0{ŭ ?t7%VڹGBke%r~r_!zÕ" >ڷVbQ\fL=<50b<K*tz.I?>5cW1sV?&C"$zL5׬8jW?f U@PHL@! Cac1 cBVY@۳u=BV9G-$ Mi2hiWqutVpErnMFBr9u uUa&|_w=\0j8d jx`XB!,2my3N=6y½q(Zx b'YON/ j^ZHymk–~Ep!~G$9VC$I,#!y{F>_.x5YW=)zxS"(%vJ J2B6& q<Ib #NL)UuI͢L9+ ! ddЀbygt1(\`MqR$RtlL߉>: G Ι@"*%M҆H"UPp 6I`{z5wr !xeݷɾZX\o,e+ A R58Q5X_w!01)܌VG?A5WBNʅ7ϖYs+ryB.e~b &+.x ק3L3*P$YR׻+uʐ (Jus{0{=]ź#V^AlV1J6Z,G,QǼb%hLnaKw9܊ wNWSz u`e~P31G9'Z䍅ca;@hc,8B '_k+/\aMp[,կɇ Gڴ]30(Ht|v2Q a\60 "eccӹx/ݡj,SsiQ+LF8tu rJ0P!VQ7ݳuD(^sO@X$@H"V" ,FNB8- J՜o_zwD$OkJ2V]4ɷ;(%CzLZkLA;,<.5de%^;WU`2Xk!wtjajzw.% Wn|x$%D= U)vKUu 90D2Oi/nPVNs]$ {=xP܅+QWa8+$v;=Fj..zvs8cQs][or+rPX &8Al%WK^]$9O5%YC{3Ñڢ]_UW׭$Ȍi#p i%L*d* ; mkE{Z`Uފб*aoxH?[llDH2=kxX F4lOQv#1+;FwKpX!8THM:TGa 9@}B14C,0@͓ Ͼ[);>v&inp%8uYy]EZps;Fvί8`Du,jhGj{4`z%]CqAMAvʰP*TRUtRSN2|ٔʠ{"n NHk *H)!PC ,4x'0.'It#lW$ƼBؐzBGC iW^APֹ{+pu9%j&96ЄLL恺"pir)nh1G3 G$I@7Ma^?'򿬼_\]v~x,f/X#၎y2^.$7y) _~5.4 ,htK-tȂLX07; *ĿQWTBԃxr(xsQ ~C(-|)-^ESLj#1StQ PM[:ğ <it&P!S!0Ƈ9PO)#zص* Dac*ڮ'ڬ=tbcux >w<]5 97>E%jgCX |BˉKNL2cA+ ָAnd7;LJu|ӵb0txӎ\ j8涳XjJJ-[Xe]eчpn$ R*7M+o?=8 7o ŰW8ǯN+CqH29uO^.US?qs8~_0 j4iC;dڼC㠼'S8BsIJZ {%&[ָ7Jaӕ}Z$CnST;BP01ڼ"HW1)bX"/{=|+Rhl.zw)qqӃp6Ozٱ+Nǯȍ6&`=^'3Rt"ecZ WV{(EHqAG!#x!-$\~%|R*S=?ZA!՘[»bn՘Aaѐ*6D(&$1fGYNTgAIeBI#̰H{aj+MHjꪈh|\JOTʰC]Jxĥ!] bAʘ9#O, u A`9 5A *Q8"%lLtG|9Tq6Q 0$h83ƍv>8oO[;|5V3:>Ѓ3v F ,STTҾPF`C SS9>PeHm $e]$HZ0p GqӃp6:JFv$I:xѣY.ގtNO j V 'HZ/5Z`'u%ҍ,}Q!Sܝ°1UL `>qdyoϯF^G IJ0OzSy}ײ9-!l z)hmG!|K[DDn >m=ľ7SgK9ty/9Rt~ UCzd OjTbTSq(h#޽m b T)GD,P}_>_I˓k$f µ9H> ?>j:nab랷l~}:˸qvO]yGBUjB}VLbj`A1ٲ}B FJdFTka_gccznqrT"ft!1Tl*߸"L{o|5SQ"% oF+˅D\DR2`8n}"ZWL (zjgt4kwG!gfH}O. v.yU$ - d6[mXoHd|-"/!}  935۸t&I+EDBxa E㒆5-RT6QɲիwS3羱_n7گ60 2 WxԣWmq-й |{Z|>G7x 3987vXA+odutx6=E K3c(Ε>|9WK!F}nIܟl>! ^`.&Yd=/Ľ-*(i" BRѓ2:w"ڰBK S"H²IRAyoAOHq]ve}%<XцE518|ݹ|73DSjW5'rbm-R[bSRJ-bٖ5^(QJе.D(}2 eshqtfqPvO:5cRK;љO$XA C^6X|Fe,*X_E [jъ؛KEzS7ŨF\aBuW-kt6J2`1'+)$/ǁL26BA $Ӕ*xɯޘ}=o(rV[hWieI:P]шgTSu2`8nx%@ ygNVxⳮWay{ ˦_X]QɌ;dw/˦s[xq3B 'RRlұw@PʡOuOQGa@ُ;#r+g 5j$yki#J'fQ/Yfggw sOnu,m^4x:w3C.LU\;U\;mV}_O|C>H1p+4K\h{5DQ uɠwۻg3<7*9]Ir3?͸cNr5Xyo,|qvr[hMڅv.).] n-@kRB( Qu(.׆JSx\{hmx54g9kiDMs&j3Qf&jŚ"v3AAE (D,M++5WD%bA[ YI+lvJGt޴i9IrOk>ͮ4ӦkiհBZ+्"F2ԚCFHImH.T%ҵ/Ҝk>.l-kFh4ZWjq|h8zrGt:ЈVSJT[P%caGbXkh8.SGJb%Z(KA)M!hI]NQ~BTBCwu6n]Eu ICX"A5PP_rm d%GULKڎ`ZcvLn`).uq~.њ;_a?IKMx3oowxxɏl8wr"k`:Oq88/}m..__~5{Y gv̿^϶ߘxa, d46JIR$B6QABsA 32y8`2N$(<^ κB5f%uw#?^_?{㶍1Ɲ d 6Anq)Hf؎yt~%#r"Y=utlI&xHPm_?&@',\\-\rdi&7?騟x//=Ov_ v?.ٟ@B{ {ƴҽ'ͳuv{p&{~_:% ʃWO_s$h4&tx|O*A];a@ 5ۗn­l>]?XrA|8ܛM<}]W8<t@ Oc.O[aEDZn*>>8&r .2yz?nr?YŇI&ƕ3_nPG!;_=\V~@o)دx6~тo!*lS0irՕ&v I뫉^ ^1Iz3 6V?mnk&K~+!unEd#/F~6wS!H7Gqgֶ66A{zzٟ@Io4ɳp8P |p;cGt]E(hr4~Lu{oOfm< FW]=&fi .kh>NqJREOi&t2s oڗW3^I9D8_ ~.|R- be.Lc,խξ߄!3IY-\mM/ϥY,da?L,Loii%AcO.nð, ?X-ˏҘLQ&Ϝw1|[~c|s"Y:ڮ:/2sV+?r=]Hw98 bpfB>K=f Y3 ,$84H *mpF8:ogo-Y09=!͝C24Պ>ef8; 5DéF,i/ H#$!Ib5fT2zk} ׆pagC5@c.;$q"JZx+経R|vcexlF@Jʼnܹ`h,^˘HO01&:}?!NH,Z K@2rL 3[ 9ưp<&ŵ|4#;Iw <5hqcXpī'X0Cb+d;b awϽwp=to޼k @1Ru2?L>_&&Q8!w^gm4H$r8tg(-"p۟.𽀅)vDKgPGlU 8=ÏfLgˁ@G`~DZT6jU;H2|p$cNv^E3 7S`obXw̖`溽cBHQFnd9ՋE\ˮ.☓|8nwv/TL Fg>M5_ffZ/f 4AGԲ"A^Ao?,T- ;- JV9:PY ҇RӢϞQDPpy(m9-4iMNk718tZiGIiO2͠qP<-4߷W޲>`V0t JJpYӲI];ȼFRuiA8G ˷n( }f;p׹#it-vJVƂ0Զ ժ XE'r;㢹իFSq$HB& CvRXZsD&kN[ѺƝҪ"S{ AlJl]!}{*tA,F8wөv~ZUEc&g/yCz{wZ; ;Q<(ץD^ y$2/I+F-1Gj96b>,T%jw0ܖD(nx=W&?hR=% C(;ܩQ*=À۠s6DfiٕuVڝ/B+$[+W$DPec.C\RxC-T*X.^5\\ %xE4MsQi0݊qnz5(D Z8i MFYn2"Quo&_*usA̻.-Fgw2'E{FwG%H]T9IB]}o9p16ƑaELzK8jBV&EcC\]\}g 8W2JxwAIM-wO:n}y«%BvSxh.KJt !;8wgR~Ya$1\:1qlj Ml! fn3gy96sՊ]@ꕋv15^(_P/11υ&Ip!ᝢAWRh9&/YHzep`?C61sN<\pOcwNĚ׭ e} ¢@ 7yI) y ٮeSΨI,2!) D !Cec ce pK`墋`>+ joEȲ *eaX#xp uV9 {b.Qb/|,:] -PҺRLB2ɹFloØ_lL7Ţ waR YydhЕGX͂Yx" D318--hbǚm1RL@_U\M 5Dbg}n%D J*I3hVY'\9r۹] M)~ m O٣u r7~Kɝ\W8 !;L e J $I"9N^'k}%`& U`F|2TːYeN⣋FwP6fWݬB/LO%6v0ln\X~I!c#|{K%J&m%TX AfwnX~ÁDH8B#cŊH!X)m6^ Fv,omyTHXQz0l;iݟT68m9Q-DE2)Z0qқ D2`-e&8 ܛz[ i e(;sR: p>CťnvqwԪQΘbjP amsE2)$/CP]}"}AkN& N- _۫r)9A*D-( wT"iU`-hS%:b'w1Sij_'  MOPA;]ϥk 2/1'7RJCu{0i"j-"HR8QKO 3J4X- d5±5BicrQx?{Fe/]D2ߏz$v :/=0hZ INwИd,,4&]R]C^^^\33NpB!)$I.3d "L+^2T[e o[):N#!Acnc#) |)RUM:KM%loU`;1PYtՅmvCC{w#@໻_ݯ\7~*h&C&*/*h> B~KD%M" Ў955گa;yY`I'G0XLPq˞EY΀ړ\2 v 2f]%{i\}͕iӧ́yߟWn{ j;^\jLVaAe3fi6A,4\CI6[?yp0yOp# ([7hEC8CΤQ,* \S=}6&ٍ:ꃘϿ͗u?Z9z+>MI )TQ K:"<Y4jb>[k5$ npPB3rk51v^?|Niyx TFDП.aCI BK9ſJ †9gĔ c"ӡȂ& m3wބʘH XhgPو&#ya5+u: V'Z7Z- LswRM(Z/P|Ω,^A1UL]OA=Hc9C8'8@bSV3!)V PBʑP2",.x,-%+f RFSJL9"R&YAH Kܪ6e)zwu~y\pX7(v7Ԣw;sER3VyfGjXHLyIpk6-2L#4Z-'۝$02L#n.b$ 4@;7 53ʺE"g.߫VܨN[OA#8 vG'a!y̆D2$8) SK.TH5y{^L^;l#O?~~aڡ"#"@h4,??Tq.攝r>Оebxxn0O^p~AHΡq'=?Wt1h"Y`ҷcxH,c+dpTcC9#=3L@NWO {!0l&StO 1%8 J WHDTe<1bI*ܧs" }'aUKf DiY? :)X\!Oe}ώR8Եox(H`,ɑ",e>Ǟtt~ 0>""-~CIcϼ B9 Ȣ"0gXCb/Q,) ,"!8 `(cI&r3cR e19Rm%jjuW6:ø~5}<fqkƖ.xcz\Uz(KOU'y l UX' g*3{G>A$EQsķψ$ئ{ޛ z$ؖ$f) eR`nYp )dZGFaFa͂C`pM8~ 7mGz#:T! nz:"eF%9L3G +:b0!p.BűR[.᲼R#<9}QoRR\gyR?5$$)B^j:+$HXCx"Gé~ 2U>L`VtA쾋7{t6+T2JΨLW6aip\EkStF!Nm pF;GhAF]_h0yijɥa ƅbRBX(sti3&MA}5:XDD.z{ް6"wDMiFoF?(ڛOz&O=Dnt6^|aHqXgKGo-0]CY}_Ct̴{;c `:{\]n3En\/_]-jwo2 F@X'dnDp޹ڔ13JPlʘ}/餌@wʗYg~˗qCHX+Vb +0w,4VCYhQlkՃq"mb\t)|K]tQ3&<}"kCD_b(e VVIA&ӁTtcӱG"lt1袢S#AsP)DY'D)!D9-2d$rYryHᎤ KX\ #MafaԢ0IHQMFSlUC-;UF::emO5CG=UҪSPҪO9(HZ=8=IZ5as-Bф6Jd[)VJ -sUi Ri!])-4!ZJ o+-TOn3\)-3JKA4 J:W< ^C^CΓWxĕi[!!AҋvG#;:\xG$' +6@k ^C/xL \⒘Lë}# 4q1ts<_-u^M@KҹW# !IXkQjݦDM!b`a&3]D"lMx ϸ!y=O(!j! #U{SbC)Fˌ)ذaA9Cku@M6@K/st_}.@4:P ݙɚ3 n7)ૹ"}r\+}Qgo1pICC]Ԫ%[BvtNFGř)ouf0%;y!lvcC "B[9 Fk 3{Ѥ!jhI.[ĿouJrL{|+-s?ƏfWWi2M09;IZv[}mkiu nC!tOzk'1 Җce8yRQUgӉEG1"CPtH8UߑY.1E4KQ]5yWy (_NQ)#⎩\}DG1+Q.#2ewi[k7 ]I@)L{ /2vdsN=ր}*8EC~#&0O XQE$)W,iS^ӬmZ]U(AFt(13磫 ᵭ!x~#Jc^hOǝ|ӧl8y/p/;b9Vdaj۫nꓤU0b0<ܸFˑd❺&}t.*5"M Pg9OeE"E#8Id,Ƙ3"4Qc҄Ͱ)%C䙀9)Cq4SKP,OhSD.&Xޭc `e}!Xj]wCLr=gqU_\b=Ux+u侖KU|>tF3&)X2}5_oo^ ^?ޱ ].;awmD^G"-ސb4Y(|<=Mgŷ O|QG$` Ƕhh4TRr9yVz/U KO&kL@^<4S}t^3QjXc~CUW?^s͞[o˂N($ E\ONn31B9vUXYf`PU;Bh`#9<>?dN5,|G=.xjxZ bmvx<fL-B K'#Dn=[v"ucRkq\` w4V7`BD$M3Q{t6+T3 nR'qO1R,W=י~+:];#YU&ο-UT"1YjyUFT#䅑&9] \`9ZIRp2ɭO /N7v_O/J/V~_á4>؇? c03ot3b̀l\KzVYQ,i*3Y.766F6vhS]^I݆ژ(^zAbEɨ SPbFB"Z\' wE%E4G?[7CM*QQljr|x^2ъ+ޓݻ$6#yьM o)~= ":GbRXo~:3%.B NzB]+{T*+yR*PF䟝E✻3xKL3xNA`<(vy'"oeH&dP|_}]%Ѡ2`7 9U.h݁HP C0 *逼 9E+HV !0TdǭL+đp,G˄afMB캗AFxX?pZ\!|j.sZ)j/b>(w~!w7{;z'mXs΄[A.!ْ7Ke(+1Mj u}`q^b=scYBGq9-r2NhZ͠HPR v@4#2߽8h6zbHaTޅX.r 9;KaEq]z{+F~U.5IaJ/FJ; ##.WO7s"HM{M)d֭Q' !ީ]Kӟ- ¸wVE~zGO9c;N+H\nWR~eP ~EP.kѓ~S >3AR:$W!.kz7VЧXHVf zqJZES(%@Ӌ8#h9Z.v%$'УC!gzX?P諈@B`!BIaa \ uy)@DK%%1%s83#)< j!BS!@pa^y`R5,Hjq'h_Ǟ_ZKM*ulSE,*%:UQp{垙c\R -r5 Ӆftk_h@݂ǽdIA=͍I?lSu+ͧa\-.7ķ nu5!D6s׉@}MĘ4b(wݦ& uUڔ98mJTFj^5{|ѽNz31NrzI.2bZĻx'2;"%$5{ZU 0__rXCkj!v9!,UJO{Q=N98\QgESa\7.vB)h.[+s0FR:O">5DUyB;O؇? c03ot3b̀l\KzVYQ,Dօ]oll@lм Vp4E 11Q"6},=Q\WELa DQD:8ߺA4FTgqxZEBݖd38=՟Owv.kw];=u;7z٤VDOFqKeb=@5 D\WLWؙX4dpb\`2L,y0p2LN߱#mSco6 HVYv]!^b<@"%~U 33D9fbZ/r6KF+:nLz>vGǛL{6[1rm!>QlD= 'z>xtH@px3s^b!lWz81Jϻ^ɠjHW7L=Otgőo=~5YQSsk =h]dH_?SNfR 9y{Y̍S2Τ (K<( uy'"P%H&fozu1JܦߢBf׼OW8Qs:]?r D{ 7%lxP!IňN p1ДނąP.k  W͕cDa$6 |>PFԴb}epsT.3lO)$jEXF-PjBĒt"g$"ajoDGzc=R)j>\ Fzi/K<*̑>as#\#8D~Ǽd&K".f$p(W\ I)"Rht:aARC֦f;QFu>L%Z*mWgyQE/!Z /L`9cBj${f 6ُ#~lhhadCdl;SHRʶ@{O6k sZfO|9Z"Q-"heSHf$Z"A${D'-w#Ɩ8LjťexҏԨMLt ~ʨJ}wWJ \g`p9 sV;Xry Kk-I5l%oKJpd.eqfCd15oCIFKP;"95v :KDK0N2][. hRқ>6Ic^XY'OEVSdM "~eToJgR&3|+oGoqn.GtFНr776 Z:\a+8q0q-WF 1DSykcyzaaQh hhT;)?y}L؎N)hBa%n @%8q-cU`IP =%@hN)`'̬e)uCXE|$ndCt˕~NU[k/9q̧j ɤKXS0!yg42ǝ.&diCeN~pW}s)`xe׭=:lV#;' HE:JTVT k? Fq):QymjqO12,)puq9ŷȪӵ1R?dFgI7/jKu*Ujc8I*hAPH #Mr2A^Yr>p+d[/^o쾞j_{^<ޛ'C3si|ƒafBgĔZE ;XSvua#;4w.\/MQnCzLmL`HM_d= ĢdE)(WU#!s-.SvEC0MϿ )doVьŒ>Nctk̐ɍVt2_d%&ɳ e℘D)&{}֛b7>b 9gZU1i}J3l0YtlfzZ8V ܹAֶ~5e[=Y7VkT|.R¾ڂKqIj?v&V-;"0S^}1T.U^I6y0pLf<~q^&أOٿ0 Y^]d9hv)x DJ1pKyL[=_NVI'&ǂdĂ[HU/[g)z^Uw2_0#utH*bc92,C8{%|fWvbe+UfL4X96H*@& xl\<36oؗʍb Y"́&=^]edBOҤD"ц:&"6pvBRVu[\@BeʀCڽmQ`)oP}j ivj !mF%d~7N$gc6"f -l\ Uq O27blوX&-ܓ%RQC|\@`_bXI ?=X6y ȄG~$Hs?d"~$^XI,v*K=,<<@I6&Rr˘!-0JĤ"CpQˆ:XH|cW%$ DbD-0QPt\(Lo2! F'dyn*@2L\RTLl<,׋g,y_åc|l3ظRuqx- [w"<WB$Gh4|~# \G`*_㩞@ >m_0A:*;zRDi}ۚx pcM&*ܜp*q"ezHIC %J ,Ѱzhp)p$ (N'*iCU>ĉʱ9^Ǭ4"$3f!ogdFś@n=>ljَjK [T5qYx|Ǎ$_OT :Om #d/Ֆ>jf~5nkEVM*12I 7NS#jǤ$F16whZ{W_sDm³f{졖観p4Q<J٨u@KOpwnO1! +@ad4走Rbn# V9 >94t`J܀ɸ@#$:8*qrB1bHG p/:BKV8G\ ]au֠nf߿upJX }`=B|TX-,6_^ߏ>ȅz3ybR+?OOnĹ?d)N9a[BSx(\Mۿpws)*}CI @…`+'<1u'_xÙ=$BcFR4>7W%Rs;Õ!TVqY h +RzNiBEΔKθ~oj#p]K\\* ?j _*)dA(-OfH2 UEl%TWI)Vƫ]IW<;!أXgUw>Ov.T$D .PN ֙t*ѝO*:6a4\lW:-Qljզ鯩^0~féf=tOݯp#ޙc $?VW5>#^l+"Soh6 |+Nsn>NWg$bl=~x-6U,s!I]B ^!7a+@ (b.He%* ,@xd!]qFcP6ҸQ鋋χHbIhaȀ#I_]f)`F8EGAEPz`Z*$V)_e#Ɍ )};BWVDiT디EG g( dtH(h)8$z1Py5h*UA0`hX#'9+J@JI]@k%/R"4ů 3A"3CNZ#(d@ej3Q/Sp>C6FP 68 >P$##Z4Z1sm-\B5DIyK=r :[ԯ7(x;ㅞEK.ھL[t+VG."z;/`;_K3ѷDi%rfcϓ=͂,hfIw&%}݃<=4SlIQ ƒ}1i'YE$ǫ3 mUDR-p-/o'g7=8a|}6,oG6 5ف[F7Áqʹw?97OKCWj |P-gc4Ծ7#,N:}gaZ^s׶|Ou6hwK.|;̀JJox{ܹmtCͅ :Q"+.}Wa!0䍒zvfgܯ(W4u?ìff +x~y~ Q}V(wNs{弑8+ÃY뇙VY7XDIB޸v)vD}n]1(hݞϛ.y9{Xg4F.3U;a4M*X'x{SEoWM Ջ ۬C|><=9LOOA8Y6<䃁syzGm'>@*]ϲ\j*7 {x)Ɂ5yny촽;`VFȾĚ{++cq5 \散NTDA֊#/'c] o9siSRi $]4)Pq]OfeK+>Yڬ/vb҃]830++I=w"Ew"^ܯElPz,EkL$^4jRqcLWᩫb$oyѿ+4J\F!&{=zk?,6MY8X(Ci@=3 t"`"6dpJ)A!Q]*nB@TD\J3ەF+@i.FP! 1XGa 1-ȭpB;ӄ'W0,B.0Eoׇˇ|?~ų6 ހ3#fX-,m)xռ3y$ƕ̩D-ɜ(K~7)%t-Nz@cWg4Q}Oe0QGcU}WV=.r֣͘GQs.%y9+z[0[ yn J9ONjP]çՋ6QL/Ꙛ*f\n6A0$E0^lMָނ0T}yc+?KK'guLC ! d{oF i"s[=¨iV;4إ|XbγsI:) ֆA{xSlڍ6PaoF8u)yZTP3GUV~ Kl.JJ>R0izT (ü73] 7b pm|FfYu_er !w-1g>8[1N3jvS^ey6GyĔ~yvmZ1N3jۖ[@m=)#)s`@~fhkВ;\"%lt(cw C>mwI7/_A3@y6دwv~;; -erRs[UؗwwtA\pݸ2igSF8~.A7ẉu.AM[d!h3 v+\L`Yb};1[r}e$Cg1=@RuGE";/6jMΥ#8Ǐ4H*r˅Nu(_Us(νubMzMD{k7{z;ZCmLz Sټ@G{Rː| }sPUX mxmiDwC.@4d!'J9ucʽA向}(M^9_)PKV/u3[`n391g|0ɘ8!ڈQSt4PxNV=xPھ SkELM]{Rbb>ZgV .C8PV3 (9b-.\#!y擫cb9wWj qtX"H9Pi͝0vƚzY%:'SU*Ѱh}$偷jCmLg']9zU[/69SvS~2!=De1w>\yƳ|c%g45u ǑB6I|[l>N+:^O"+3IDK9f TсZz,3߉p꾼`Lt35,Unf)XTl4,w0~o ۯ+8h$Yu y٭Q4>*mWt/ݚimm)_qg7G֊H"U*f"멘5sJUf;ѓ"aL ޜ6vco׎IѫUVpvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003405606715144732444017720 0ustar rootrootFeb 17 00:05:25 crc systemd[1]: Starting Kubernetes Kubelet... Feb 17 00:05:25 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:25 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:26 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 00:05:26 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 17 00:05:26 crc kubenswrapper[4690]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 00:05:26 crc kubenswrapper[4690]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 17 00:05:26 crc kubenswrapper[4690]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 00:05:26 crc kubenswrapper[4690]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 00:05:26 crc kubenswrapper[4690]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 17 00:05:26 crc kubenswrapper[4690]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.716903 4690 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720449 4690 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720472 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720479 4690 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720485 4690 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720491 4690 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720497 4690 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720503 4690 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720510 4690 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720518 4690 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720527 4690 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720532 4690 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720538 4690 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720544 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720570 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720575 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720581 4690 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720587 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720606 4690 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720614 4690 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720621 4690 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720628 4690 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720634 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720641 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720647 4690 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720653 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720659 4690 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720664 4690 feature_gate.go:330] unrecognized feature gate: Example Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720670 4690 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720676 4690 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720681 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720686 4690 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720691 4690 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720697 4690 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720706 4690 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720713 4690 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720720 4690 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720725 4690 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720732 4690 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720738 4690 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720743 4690 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720749 4690 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720754 4690 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720759 4690 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720765 4690 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720769 4690 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720775 4690 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720780 4690 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720785 4690 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720790 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720795 4690 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720800 4690 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720806 4690 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720811 4690 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720817 4690 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720824 4690 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720830 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720835 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720842 4690 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720847 4690 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720853 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720858 4690 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720863 4690 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720868 4690 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720873 4690 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720878 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720883 4690 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720888 4690 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720893 4690 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720899 4690 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720904 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.720909 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.721677 4690 flags.go:64] FLAG: --address="0.0.0.0" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.721694 4690 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722187 4690 flags.go:64] FLAG: --anonymous-auth="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722197 4690 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722206 4690 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722212 4690 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722222 4690 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722231 4690 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722237 4690 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722243 4690 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722250 4690 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722256 4690 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722262 4690 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722268 4690 flags.go:64] FLAG: --cgroup-root="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722275 4690 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722281 4690 flags.go:64] FLAG: --client-ca-file="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722286 4690 flags.go:64] FLAG: --cloud-config="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722294 4690 flags.go:64] FLAG: --cloud-provider="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722299 4690 flags.go:64] FLAG: --cluster-dns="[]" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722306 4690 flags.go:64] FLAG: --cluster-domain="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722312 4690 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722318 4690 flags.go:64] FLAG: --config-dir="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722324 4690 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722330 4690 flags.go:64] FLAG: --container-log-max-files="5" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722338 4690 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722345 4690 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722351 4690 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722375 4690 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722382 4690 flags.go:64] FLAG: --contention-profiling="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722388 4690 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722395 4690 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722401 4690 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722406 4690 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722414 4690 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722421 4690 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722427 4690 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722433 4690 flags.go:64] FLAG: --enable-load-reader="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722439 4690 flags.go:64] FLAG: --enable-server="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722445 4690 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722453 4690 flags.go:64] FLAG: --event-burst="100" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722460 4690 flags.go:64] FLAG: --event-qps="50" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722467 4690 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722474 4690 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722480 4690 flags.go:64] FLAG: --eviction-hard="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722487 4690 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722493 4690 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722499 4690 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722505 4690 flags.go:64] FLAG: --eviction-soft="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722511 4690 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722518 4690 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722524 4690 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722531 4690 flags.go:64] FLAG: --experimental-mounter-path="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722537 4690 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722543 4690 flags.go:64] FLAG: --fail-swap-on="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722549 4690 flags.go:64] FLAG: --feature-gates="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722557 4690 flags.go:64] FLAG: --file-check-frequency="20s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722563 4690 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722569 4690 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722576 4690 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722582 4690 flags.go:64] FLAG: --healthz-port="10248" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722588 4690 flags.go:64] FLAG: --help="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722595 4690 flags.go:64] FLAG: --hostname-override="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722600 4690 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722606 4690 flags.go:64] FLAG: --http-check-frequency="20s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722613 4690 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722620 4690 flags.go:64] FLAG: --image-credential-provider-config="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722625 4690 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722632 4690 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722638 4690 flags.go:64] FLAG: --image-service-endpoint="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722644 4690 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722650 4690 flags.go:64] FLAG: --kube-api-burst="100" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722656 4690 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722663 4690 flags.go:64] FLAG: --kube-api-qps="50" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722669 4690 flags.go:64] FLAG: --kube-reserved="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722675 4690 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722681 4690 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722688 4690 flags.go:64] FLAG: --kubelet-cgroups="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722694 4690 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722772 4690 flags.go:64] FLAG: --lock-file="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722870 4690 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722878 4690 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.722885 4690 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.723859 4690 flags.go:64] FLAG: --log-json-split-stream="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.723904 4690 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.723920 4690 flags.go:64] FLAG: --log-text-split-stream="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.723932 4690 flags.go:64] FLAG: --logging-format="text" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.723946 4690 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.723964 4690 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.723977 4690 flags.go:64] FLAG: --manifest-url="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.723990 4690 flags.go:64] FLAG: --manifest-url-header="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.724857 4690 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.724875 4690 flags.go:64] FLAG: --max-open-files="1000000" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.724890 4690 flags.go:64] FLAG: --max-pods="110" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.724901 4690 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.724912 4690 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.724922 4690 flags.go:64] FLAG: --memory-manager-policy="None" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.724932 4690 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.724999 4690 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725282 4690 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725294 4690 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725409 4690 flags.go:64] FLAG: --node-status-max-images="50" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725422 4690 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725433 4690 flags.go:64] FLAG: --oom-score-adj="-999" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725443 4690 flags.go:64] FLAG: --pod-cidr="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725453 4690 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725469 4690 flags.go:64] FLAG: --pod-manifest-path="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725479 4690 flags.go:64] FLAG: --pod-max-pids="-1" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725488 4690 flags.go:64] FLAG: --pods-per-core="0" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725497 4690 flags.go:64] FLAG: --port="10250" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725506 4690 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725516 4690 flags.go:64] FLAG: --provider-id="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725525 4690 flags.go:64] FLAG: --qos-reserved="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725535 4690 flags.go:64] FLAG: --read-only-port="10255" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725545 4690 flags.go:64] FLAG: --register-node="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725554 4690 flags.go:64] FLAG: --register-schedulable="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725564 4690 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725582 4690 flags.go:64] FLAG: --registry-burst="10" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725591 4690 flags.go:64] FLAG: --registry-qps="5" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725600 4690 flags.go:64] FLAG: --reserved-cpus="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725609 4690 flags.go:64] FLAG: --reserved-memory="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725622 4690 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725631 4690 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725641 4690 flags.go:64] FLAG: --rotate-certificates="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725650 4690 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725658 4690 flags.go:64] FLAG: --runonce="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725668 4690 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725678 4690 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725690 4690 flags.go:64] FLAG: --seccomp-default="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725700 4690 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725709 4690 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725719 4690 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725728 4690 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725737 4690 flags.go:64] FLAG: --storage-driver-password="root" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725746 4690 flags.go:64] FLAG: --storage-driver-secure="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725755 4690 flags.go:64] FLAG: --storage-driver-table="stats" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725764 4690 flags.go:64] FLAG: --storage-driver-user="root" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725773 4690 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725783 4690 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725792 4690 flags.go:64] FLAG: --system-cgroups="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725800 4690 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725816 4690 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725825 4690 flags.go:64] FLAG: --tls-cert-file="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725834 4690 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725850 4690 flags.go:64] FLAG: --tls-min-version="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725860 4690 flags.go:64] FLAG: --tls-private-key-file="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725869 4690 flags.go:64] FLAG: --topology-manager-policy="none" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725878 4690 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725887 4690 flags.go:64] FLAG: --topology-manager-scope="container" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725897 4690 flags.go:64] FLAG: --v="2" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725909 4690 flags.go:64] FLAG: --version="false" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725921 4690 flags.go:64] FLAG: --vmodule="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725932 4690 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.725942 4690 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726171 4690 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726182 4690 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726190 4690 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726200 4690 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726208 4690 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726217 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726226 4690 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726234 4690 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726242 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726250 4690 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726257 4690 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726268 4690 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726277 4690 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726286 4690 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726294 4690 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726302 4690 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726310 4690 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726317 4690 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726325 4690 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726333 4690 feature_gate.go:330] unrecognized feature gate: Example Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726340 4690 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726348 4690 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726398 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726407 4690 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726415 4690 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726423 4690 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726431 4690 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726440 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726448 4690 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726457 4690 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726464 4690 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726472 4690 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726479 4690 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726487 4690 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726502 4690 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726511 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726519 4690 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726526 4690 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726534 4690 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726544 4690 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726556 4690 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726565 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726573 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726581 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726589 4690 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726597 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726604 4690 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726612 4690 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726620 4690 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726627 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726638 4690 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726647 4690 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726655 4690 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726663 4690 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726670 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726678 4690 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726686 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726694 4690 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726702 4690 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726710 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726718 4690 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726729 4690 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726739 4690 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726747 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726756 4690 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726764 4690 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726776 4690 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726784 4690 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726791 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726798 4690 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.726806 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.727688 4690 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.739572 4690 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.739614 4690 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739727 4690 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739739 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739748 4690 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739756 4690 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739764 4690 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739773 4690 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739780 4690 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739789 4690 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739798 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739805 4690 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739813 4690 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739821 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739828 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739836 4690 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739843 4690 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739851 4690 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739859 4690 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739867 4690 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739874 4690 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739882 4690 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739890 4690 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739898 4690 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739905 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739913 4690 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739920 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739928 4690 feature_gate.go:330] unrecognized feature gate: Example Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739936 4690 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739944 4690 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739951 4690 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739959 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739966 4690 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739974 4690 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739981 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739989 4690 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.739998 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740006 4690 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740013 4690 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740022 4690 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740032 4690 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740041 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740050 4690 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740057 4690 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740068 4690 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740079 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740087 4690 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740095 4690 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740103 4690 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740111 4690 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740119 4690 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740127 4690 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740135 4690 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740142 4690 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740150 4690 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740158 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740168 4690 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740179 4690 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740188 4690 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740198 4690 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740207 4690 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740216 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740225 4690 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740235 4690 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740245 4690 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740254 4690 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740263 4690 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740272 4690 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740281 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740289 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740298 4690 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740307 4690 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740317 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.740329 4690 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740617 4690 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740634 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740644 4690 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740653 4690 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740661 4690 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740670 4690 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740679 4690 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740687 4690 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740698 4690 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740709 4690 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740720 4690 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740731 4690 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740741 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740751 4690 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740760 4690 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740769 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740777 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740787 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740797 4690 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740806 4690 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740814 4690 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740824 4690 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740834 4690 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740842 4690 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740852 4690 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740860 4690 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740868 4690 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740876 4690 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740884 4690 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740892 4690 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740900 4690 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740907 4690 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740916 4690 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740923 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740932 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740940 4690 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740947 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740956 4690 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740965 4690 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740974 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740982 4690 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740990 4690 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.740998 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741006 4690 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741014 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741021 4690 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741029 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741036 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741044 4690 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741052 4690 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741059 4690 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741067 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741074 4690 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741082 4690 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741090 4690 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741098 4690 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741107 4690 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741114 4690 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741122 4690 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741130 4690 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741137 4690 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741146 4690 feature_gate.go:330] unrecognized feature gate: Example Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741153 4690 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741162 4690 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741170 4690 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741178 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741185 4690 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741193 4690 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741201 4690 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741209 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.741218 4690 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.741230 4690 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.742482 4690 server.go:940] "Client rotation is on, will bootstrap in background" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.748167 4690 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.748310 4690 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.750488 4690 server.go:997] "Starting client certificate rotation" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.750536 4690 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.753465 4690 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-09 20:44:47.418874571 +0000 UTC Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.753579 4690 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.776896 4690 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.780700 4690 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 17 00:05:26 crc kubenswrapper[4690]: E0217 00:05:26.781505 4690 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.802320 4690 log.go:25] "Validated CRI v1 runtime API" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.839517 4690 log.go:25] "Validated CRI v1 image API" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.842171 4690 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.849288 4690 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-17-00-00-51-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.849336 4690 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.870980 4690 manager.go:217] Machine: {Timestamp:2026-02-17 00:05:26.867976453 +0000 UTC m=+0.603689234 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:ed6e37f0-bc2a-4f72-8a62-fab92258d3ce BootID:e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:61:cd:c2 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:61:cd:c2 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:7e:65:23 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:cd:fb:a8 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:d8:01:57 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:8c:a0:5c Speed:-1 Mtu:1496} {Name:eth10 MacAddress:fa:8c:cb:0b:05:68 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:92:0e:b7:23:bb:d8 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.871243 4690 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.871426 4690 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.873252 4690 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.873568 4690 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.873612 4690 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.873870 4690 topology_manager.go:138] "Creating topology manager with none policy" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.873883 4690 container_manager_linux.go:303] "Creating device plugin manager" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.874404 4690 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.874444 4690 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.874675 4690 state_mem.go:36] "Initialized new in-memory state store" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.874779 4690 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.880244 4690 kubelet.go:418] "Attempting to sync node with API server" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.880269 4690 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.880385 4690 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.880399 4690 kubelet.go:324] "Adding apiserver pod source" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.880415 4690 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.885730 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.885788 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:26 crc kubenswrapper[4690]: E0217 00:05:26.885854 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:26 crc kubenswrapper[4690]: E0217 00:05:26.885871 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.888916 4690 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.890102 4690 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.894068 4690 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895694 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895732 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895743 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895753 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895770 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895781 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895792 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895809 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895824 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895834 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895851 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.895863 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.898394 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.899074 4690 server.go:1280] "Started kubelet" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.900527 4690 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.900839 4690 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.900839 4690 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 17 00:05:26 crc systemd[1]: Started Kubernetes Kubelet. Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.901548 4690 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.901926 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.901961 4690 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.902079 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 00:54:01.26443489 +0000 UTC Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.902226 4690 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.902240 4690 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.902408 4690 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 17 00:05:26 crc kubenswrapper[4690]: E0217 00:05:26.902652 4690 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.902854 4690 server.go:460] "Adding debug handlers to kubelet server" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.910624 4690 factory.go:55] Registering systemd factory Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.910692 4690 factory.go:221] Registration of the systemd container factory successfully Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.911504 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:26 crc kubenswrapper[4690]: E0217 00:05:26.911786 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.913868 4690 factory.go:153] Registering CRI-O factory Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.913905 4690 factory.go:221] Registration of the crio container factory successfully Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.914016 4690 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.914057 4690 factory.go:103] Registering Raw factory Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.914088 4690 manager.go:1196] Started watching for new ooms in manager Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.914753 4690 manager.go:319] Starting recovery of all containers Feb 17 00:05:26 crc kubenswrapper[4690]: E0217 00:05:26.914730 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="200ms" Feb 17 00:05:26 crc kubenswrapper[4690]: E0217 00:05:26.911944 4690 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.181:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1894dfe4b82e66a1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-17 00:05:26.899033761 +0000 UTC m=+0.634746522,LastTimestamp:2026-02-17 00:05:26.899033761 +0000 UTC m=+0.634746522,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927325 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927445 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927470 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927551 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927575 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927596 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927618 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927643 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927665 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927687 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927712 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927732 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927752 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927779 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927800 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927836 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927862 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927880 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927901 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927924 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927944 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927964 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.927983 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928004 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928028 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928047 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928130 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928152 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928170 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928190 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928209 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928230 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928250 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928269 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928287 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928307 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928326 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928343 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928383 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928403 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928427 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928452 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928473 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928492 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928516 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928535 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928556 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928580 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928602 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928625 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928645 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928667 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928695 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928714 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928739 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928766 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928786 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928807 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928824 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928847 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928870 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928891 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928913 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928936 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928956 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928974 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.928995 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929017 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929036 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929056 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929074 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929090 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929109 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929132 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929153 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929175 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929196 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929219 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929239 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929260 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929281 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929301 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929319 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929342 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929388 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929409 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929431 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929453 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929474 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929495 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929516 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929530 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929545 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929567 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929590 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929617 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929638 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929654 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929677 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929696 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929716 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929735 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929756 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929775 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929803 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929825 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929847 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929868 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929889 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929910 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929933 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929955 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929976 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.929997 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930017 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930038 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930057 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930078 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930101 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930124 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930144 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930163 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930182 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930200 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930224 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930242 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930263 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930288 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930310 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930329 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930346 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930442 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930466 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930488 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930508 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930528 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930546 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930566 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930588 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930607 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930627 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930647 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930666 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930691 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930712 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930735 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930754 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930775 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930793 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930810 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930832 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930851 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930874 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930899 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930922 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930945 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930969 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.930988 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931009 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931053 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931075 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931094 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931113 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931131 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931153 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931171 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931189 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931210 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931230 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931251 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931271 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931291 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931310 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931330 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931349 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931396 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931415 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.931435 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934191 4690 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934239 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934262 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934282 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934305 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934323 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934342 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934382 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934399 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934416 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934432 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934451 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934468 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934489 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934506 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934526 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934543 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934560 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934580 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934597 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934617 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934635 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934654 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934673 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934693 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934729 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934749 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934769 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934788 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934810 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934826 4690 reconstruct.go:97] "Volume reconstruction finished" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.934840 4690 reconciler.go:26] "Reconciler: start to sync state" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.944904 4690 manager.go:324] Recovery completed Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.958394 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.959992 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.960065 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.960091 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.961562 4690 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.961591 4690 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.961623 4690 state_mem.go:36] "Initialized new in-memory state store" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.969512 4690 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.972923 4690 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.974635 4690 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.974738 4690 kubelet.go:2335] "Starting kubelet main sync loop" Feb 17 00:05:26 crc kubenswrapper[4690]: E0217 00:05:26.974925 4690 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.975885 4690 policy_none.go:49] "None policy: Start" Feb 17 00:05:26 crc kubenswrapper[4690]: W0217 00:05:26.977108 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:26 crc kubenswrapper[4690]: E0217 00:05:26.977180 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.977924 4690 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 17 00:05:26 crc kubenswrapper[4690]: I0217 00:05:26.977956 4690 state_mem.go:35] "Initializing new in-memory state store" Feb 17 00:05:27 crc kubenswrapper[4690]: E0217 00:05:27.003692 4690 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.044583 4690 manager.go:334] "Starting Device Plugin manager" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.044690 4690 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.044707 4690 server.go:79] "Starting device plugin registration server" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.045197 4690 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.045215 4690 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.045834 4690 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.045935 4690 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.045951 4690 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 17 00:05:27 crc kubenswrapper[4690]: E0217 00:05:27.056172 4690 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.075435 4690 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.075610 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.076907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.077008 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.077084 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.077385 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.077529 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.077590 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.079011 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.079111 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.079197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.079152 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.079377 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.079388 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.079573 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.079754 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.079812 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.080496 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.080530 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.080539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.080644 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.080842 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.080893 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.081220 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.081307 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.081335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.081417 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.081430 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.081455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.081543 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.081690 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.081723 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.082690 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.082697 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.082726 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.082745 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.082746 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.082708 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.082783 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.082762 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.082810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.082983 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.083009 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.083608 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.083632 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.083642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: E0217 00:05:27.116647 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="400ms" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.138572 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.138616 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.138642 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.138660 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.138678 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.138698 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.139078 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.139330 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.140291 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.140764 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.141334 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.141389 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.141444 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.141518 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.141551 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.145675 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.146933 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.147059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.147083 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.147155 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 00:05:27 crc kubenswrapper[4690]: E0217 00:05:27.148023 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.181:6443: connect: connection refused" node="crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.242714 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.242789 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.242819 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.242845 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.242865 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.242888 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.242923 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.242953 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.242980 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243002 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243060 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243079 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243100 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243121 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243141 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243145 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243205 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243262 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243219 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243324 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243155 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243407 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243131 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243411 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243452 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243459 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243497 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243532 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243584 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.243653 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.349047 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.351595 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.351637 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.351648 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.351673 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 00:05:27 crc kubenswrapper[4690]: E0217 00:05:27.352567 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.181:6443: connect: connection refused" node="crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.413478 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.437614 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.456425 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: W0217 00:05:27.460011 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-c64f5af11ef65dafb23db09ad46cbc5ff9545557b723769bc6ec3d036835b9bd WatchSource:0}: Error finding container c64f5af11ef65dafb23db09ad46cbc5ff9545557b723769bc6ec3d036835b9bd: Status 404 returned error can't find the container with id c64f5af11ef65dafb23db09ad46cbc5ff9545557b723769bc6ec3d036835b9bd Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.464473 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.470473 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 17 00:05:27 crc kubenswrapper[4690]: W0217 00:05:27.479485 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-0808852d5d484bc4126ed4589c860a501c76ae584f56f8df45f2878a5c80d3fb WatchSource:0}: Error finding container 0808852d5d484bc4126ed4589c860a501c76ae584f56f8df45f2878a5c80d3fb: Status 404 returned error can't find the container with id 0808852d5d484bc4126ed4589c860a501c76ae584f56f8df45f2878a5c80d3fb Feb 17 00:05:27 crc kubenswrapper[4690]: W0217 00:05:27.483464 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-9d6e406074379207ac91050a117dae18eb282a18cf8d52347c1ce81e611528cf WatchSource:0}: Error finding container 9d6e406074379207ac91050a117dae18eb282a18cf8d52347c1ce81e611528cf: Status 404 returned error can't find the container with id 9d6e406074379207ac91050a117dae18eb282a18cf8d52347c1ce81e611528cf Feb 17 00:05:27 crc kubenswrapper[4690]: W0217 00:05:27.493043 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a1d04027ec92a9502f8a7d4927bc7c46519152ade1ccd4af365e6f60c08e358f WatchSource:0}: Error finding container a1d04027ec92a9502f8a7d4927bc7c46519152ade1ccd4af365e6f60c08e358f: Status 404 returned error can't find the container with id a1d04027ec92a9502f8a7d4927bc7c46519152ade1ccd4af365e6f60c08e358f Feb 17 00:05:27 crc kubenswrapper[4690]: W0217 00:05:27.498684 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-64f4d042138ea1304f3f6f3002fc8013ea3b956575c13a4d1f21c5c76a0e1d3b WatchSource:0}: Error finding container 64f4d042138ea1304f3f6f3002fc8013ea3b956575c13a4d1f21c5c76a0e1d3b: Status 404 returned error can't find the container with id 64f4d042138ea1304f3f6f3002fc8013ea3b956575c13a4d1f21c5c76a0e1d3b Feb 17 00:05:27 crc kubenswrapper[4690]: E0217 00:05:27.521003 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="800ms" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.752989 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.754813 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.754869 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.754883 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.754916 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 00:05:27 crc kubenswrapper[4690]: E0217 00:05:27.755262 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.181:6443: connect: connection refused" node="crc" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.902211 4690 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.902191 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 21:08:53.587923814 +0000 UTC Feb 17 00:05:27 crc kubenswrapper[4690]: W0217 00:05:27.971842 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:27 crc kubenswrapper[4690]: E0217 00:05:27.971944 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.980438 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"64f4d042138ea1304f3f6f3002fc8013ea3b956575c13a4d1f21c5c76a0e1d3b"} Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.981538 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a1d04027ec92a9502f8a7d4927bc7c46519152ade1ccd4af365e6f60c08e358f"} Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.982454 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9d6e406074379207ac91050a117dae18eb282a18cf8d52347c1ce81e611528cf"} Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.984175 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0808852d5d484bc4126ed4589c860a501c76ae584f56f8df45f2878a5c80d3fb"} Feb 17 00:05:27 crc kubenswrapper[4690]: I0217 00:05:27.985669 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c64f5af11ef65dafb23db09ad46cbc5ff9545557b723769bc6ec3d036835b9bd"} Feb 17 00:05:28 crc kubenswrapper[4690]: W0217 00:05:28.103888 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:28 crc kubenswrapper[4690]: E0217 00:05:28.104019 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:28 crc kubenswrapper[4690]: W0217 00:05:28.104108 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:28 crc kubenswrapper[4690]: E0217 00:05:28.104255 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:28 crc kubenswrapper[4690]: E0217 00:05:28.322740 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="1.6s" Feb 17 00:05:28 crc kubenswrapper[4690]: W0217 00:05:28.398885 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:28 crc kubenswrapper[4690]: E0217 00:05:28.398989 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.556434 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.558206 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.558274 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.558297 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.558343 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 00:05:28 crc kubenswrapper[4690]: E0217 00:05:28.559380 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.181:6443: connect: connection refused" node="crc" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.902039 4690 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.904214 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 15:11:25.41020403 +0000 UTC Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.949986 4690 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 17 00:05:28 crc kubenswrapper[4690]: E0217 00:05:28.951858 4690 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.992719 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b" exitCode=0 Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.992872 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b"} Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.993041 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.994456 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.994507 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.994524 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.995348 4690 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f3895b844cf90fffb974f0e328d15c64e1ce6bcf2e5cabc3997201315ca8289a" exitCode=0 Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.995451 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f3895b844cf90fffb974f0e328d15c64e1ce6bcf2e5cabc3997201315ca8289a"} Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.995627 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.996641 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.996850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.996896 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.996920 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.997798 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.997838 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.997855 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.998213 4690 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66" exitCode=0 Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.998313 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66"} Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.998467 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.999858 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.999901 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:28 crc kubenswrapper[4690]: I0217 00:05:28.999925 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.002845 4690 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251" exitCode=0 Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.002974 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251"} Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.003072 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.004396 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.004440 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.004462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.008785 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d"} Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.008837 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7"} Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.008864 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2"} Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.904668 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 17:04:56.934715595 +0000 UTC Feb 17 00:05:29 crc kubenswrapper[4690]: E0217 00:05:29.969765 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="3.2s" Feb 17 00:05:29 crc kubenswrapper[4690]: I0217 00:05:29.969845 4690 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.025600 4690 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ced2a900f58d82b450f0f25d729563ff1fd783ce1ee74f94cbcdb752d235ae76" exitCode=0 Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.025664 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ced2a900f58d82b450f0f25d729563ff1fd783ce1ee74f94cbcdb752d235ae76"} Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.025787 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.027923 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.027964 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.027976 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.031509 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fa8689631e853adf44aaf69172bc5b39bb6a3246bc8c668516ce10a57e1d81b3"} Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.031588 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.034139 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.034164 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.034173 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:30 crc kubenswrapper[4690]: W0217 00:05:30.042565 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:30 crc kubenswrapper[4690]: E0217 00:05:30.042662 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.043316 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9d468f64b7ca19373b20612c77a5482144dffdcd3e76d0377af6a421a4770d0e"} Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.043396 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"dae842d1ef134d705b87402029edb7e1acc4d3b1efbc458eb8ceaf07f2426db7"} Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.043415 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"971fe3f679431845df0be1a451c36f170cd038649d687a730a6b74adfe370b39"} Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.043677 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.045175 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.045215 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.045230 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.046963 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f"} Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.047111 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.049234 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.049274 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.049291 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.052947 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b"} Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.052996 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d"} Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.053019 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3"} Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.053038 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d"} Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.160217 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.161448 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.161509 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.161522 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.161560 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 00:05:30 crc kubenswrapper[4690]: E0217 00:05:30.162162 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.181:6443: connect: connection refused" node="crc" Feb 17 00:05:30 crc kubenswrapper[4690]: W0217 00:05:30.341082 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Feb 17 00:05:30 crc kubenswrapper[4690]: E0217 00:05:30.341182 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Feb 17 00:05:30 crc kubenswrapper[4690]: I0217 00:05:30.905174 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 05:48:41.227601278 +0000 UTC Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.060019 4690 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3f1b6e538bb6bf2ca276d357fae177c5a103541eb610a18cca2ecfce6078cae7" exitCode=0 Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.060134 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3f1b6e538bb6bf2ca276d357fae177c5a103541eb610a18cca2ecfce6078cae7"} Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.060192 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.061798 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.061839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.061851 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.067307 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.067297 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764"} Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.067351 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.067458 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.067475 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.067576 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.068853 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.068908 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.068928 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.069264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.069311 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.069330 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.069430 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.069453 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.069462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.070029 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.070218 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.070422 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.360750 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.389512 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.576641 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.712413 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.720817 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:31 crc kubenswrapper[4690]: I0217 00:05:31.906035 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 03:51:32.056678357 +0000 UTC Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.076119 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"85bae26ae90f236917810f277dedea3283d1ceb530696a3aa5036ce1cc1ac297"} Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.076195 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e53310be1ede7135c52d6f9f9c3a996a42129a3bfa062bd0145038944da246ca"} Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.076221 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2fa08526608473b932b6638485c2291690004e13a0e409a1dde28e531a89a166"} Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.076228 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.076238 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.076240 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2297b4caaf656677fc3d5011bd283d0e861e916f76e9e4e92e8ca12a4c211e9c"} Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.076465 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.077475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.077529 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.077552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.078451 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.078498 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.078511 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.549963 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.550242 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.551906 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.551963 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.551987 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.624864 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:32 crc kubenswrapper[4690]: I0217 00:05:32.906812 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 02:44:30.127442005 +0000 UTC Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.083766 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1f4b80ca314b3a59b982376a4fb6c98a850a7ef5a2cd1117387e414d65b61d76"} Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.083856 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.083880 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.083938 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.085404 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.085435 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.085445 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.085489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.085537 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.085554 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.087624 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.087662 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.087683 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.158174 4690 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.362304 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.363921 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.363993 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.364016 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.364060 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.448705 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:33 crc kubenswrapper[4690]: I0217 00:05:33.907496 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 06:32:55.606243036 +0000 UTC Feb 17 00:05:34 crc kubenswrapper[4690]: I0217 00:05:34.086494 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:34 crc kubenswrapper[4690]: I0217 00:05:34.087345 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:34 crc kubenswrapper[4690]: I0217 00:05:34.087629 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:34 crc kubenswrapper[4690]: I0217 00:05:34.087666 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:34 crc kubenswrapper[4690]: I0217 00:05:34.087678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:34 crc kubenswrapper[4690]: I0217 00:05:34.088560 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:34 crc kubenswrapper[4690]: I0217 00:05:34.088613 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:34 crc kubenswrapper[4690]: I0217 00:05:34.088635 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:34 crc kubenswrapper[4690]: I0217 00:05:34.908735 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 02:08:29.50680773 +0000 UTC Feb 17 00:05:35 crc kubenswrapper[4690]: I0217 00:05:35.093645 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:35 crc kubenswrapper[4690]: I0217 00:05:35.095000 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:35 crc kubenswrapper[4690]: I0217 00:05:35.095064 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:35 crc kubenswrapper[4690]: I0217 00:05:35.095083 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:35 crc kubenswrapper[4690]: I0217 00:05:35.625298 4690 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 17 00:05:35 crc kubenswrapper[4690]: I0217 00:05:35.625445 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 17 00:05:35 crc kubenswrapper[4690]: I0217 00:05:35.909090 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 15:51:41.951425573 +0000 UTC Feb 17 00:05:36 crc kubenswrapper[4690]: I0217 00:05:36.909901 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 13:58:05.125162968 +0000 UTC Feb 17 00:05:37 crc kubenswrapper[4690]: E0217 00:05:37.057438 4690 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 17 00:05:37 crc kubenswrapper[4690]: I0217 00:05:37.428635 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 17 00:05:37 crc kubenswrapper[4690]: I0217 00:05:37.428852 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:37 crc kubenswrapper[4690]: I0217 00:05:37.430270 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:37 crc kubenswrapper[4690]: I0217 00:05:37.430305 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:37 crc kubenswrapper[4690]: I0217 00:05:37.430316 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:37 crc kubenswrapper[4690]: I0217 00:05:37.910662 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-06 21:00:15.436541331 +0000 UTC Feb 17 00:05:38 crc kubenswrapper[4690]: I0217 00:05:38.607139 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:38 crc kubenswrapper[4690]: I0217 00:05:38.607321 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:38 crc kubenswrapper[4690]: I0217 00:05:38.608825 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:38 crc kubenswrapper[4690]: I0217 00:05:38.608875 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:38 crc kubenswrapper[4690]: I0217 00:05:38.608892 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:38 crc kubenswrapper[4690]: I0217 00:05:38.911155 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 21:32:10.114825352 +0000 UTC Feb 17 00:05:39 crc kubenswrapper[4690]: I0217 00:05:39.911335 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 16:50:36.281839669 +0000 UTC Feb 17 00:05:40 crc kubenswrapper[4690]: I0217 00:05:40.903147 4690 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 17 00:05:40 crc kubenswrapper[4690]: I0217 00:05:40.912434 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 11:57:11.186523261 +0000 UTC Feb 17 00:05:41 crc kubenswrapper[4690]: W0217 00:05:41.234200 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.234347 4690 trace.go:236] Trace[1541045327]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Feb-2026 00:05:31.232) (total time: 10001ms): Feb 17 00:05:41 crc kubenswrapper[4690]: Trace[1541045327]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (00:05:41.234) Feb 17 00:05:41 crc kubenswrapper[4690]: Trace[1541045327]: [10.001725432s] [10.001725432s] END Feb 17 00:05:41 crc kubenswrapper[4690]: E0217 00:05:41.234419 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 17 00:05:41 crc kubenswrapper[4690]: W0217 00:05:41.267192 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.267299 4690 trace.go:236] Trace[464365422]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Feb-2026 00:05:31.265) (total time: 10001ms): Feb 17 00:05:41 crc kubenswrapper[4690]: Trace[464365422]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (00:05:41.267) Feb 17 00:05:41 crc kubenswrapper[4690]: Trace[464365422]: [10.001881006s] [10.001881006s] END Feb 17 00:05:41 crc kubenswrapper[4690]: E0217 00:05:41.267331 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.361349 4690 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.361474 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.645409 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.645640 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.646152 4690 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.646237 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.646961 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.647001 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.647017 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.721493 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 17 00:05:41 crc kubenswrapper[4690]: I0217 00:05:41.913201 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 14:00:18.909075038 +0000 UTC Feb 17 00:05:42 crc kubenswrapper[4690]: I0217 00:05:42.112459 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:42 crc kubenswrapper[4690]: I0217 00:05:42.113885 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:42 crc kubenswrapper[4690]: I0217 00:05:42.113926 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:42 crc kubenswrapper[4690]: I0217 00:05:42.113935 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:42 crc kubenswrapper[4690]: I0217 00:05:42.143019 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 17 00:05:42 crc kubenswrapper[4690]: I0217 00:05:42.913791 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 14:53:27.942099141 +0000 UTC Feb 17 00:05:43 crc kubenswrapper[4690]: I0217 00:05:43.114938 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:43 crc kubenswrapper[4690]: I0217 00:05:43.116156 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:43 crc kubenswrapper[4690]: I0217 00:05:43.116212 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:43 crc kubenswrapper[4690]: I0217 00:05:43.116231 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:43 crc kubenswrapper[4690]: I0217 00:05:43.913923 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 22:31:22.430231081 +0000 UTC Feb 17 00:05:44 crc kubenswrapper[4690]: I0217 00:05:44.914494 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 10:41:01.760725537 +0000 UTC Feb 17 00:05:45 crc kubenswrapper[4690]: I0217 00:05:45.625732 4690 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 17 00:05:45 crc kubenswrapper[4690]: I0217 00:05:45.625811 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 17 00:05:45 crc kubenswrapper[4690]: I0217 00:05:45.915515 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 17:42:00.474547144 +0000 UTC Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.369780 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.370032 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.371747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.371960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.371970 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.378427 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.390709 4690 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.515504 4690 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 17 00:05:46 crc kubenswrapper[4690]: E0217 00:05:46.648882 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.651770 4690 trace.go:236] Trace[486363700]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Feb-2026 00:05:36.303) (total time: 10347ms): Feb 17 00:05:46 crc kubenswrapper[4690]: Trace[486363700]: ---"Objects listed" error: 10347ms (00:05:46.651) Feb 17 00:05:46 crc kubenswrapper[4690]: Trace[486363700]: [10.34786729s] [10.34786729s] END Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.651803 4690 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.654578 4690 trace.go:236] Trace[1558381768]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Feb-2026 00:05:33.719) (total time: 12935ms): Feb 17 00:05:46 crc kubenswrapper[4690]: Trace[1558381768]: ---"Objects listed" error: 12935ms (00:05:46.654) Feb 17 00:05:46 crc kubenswrapper[4690]: Trace[1558381768]: [12.935347863s] [12.935347863s] END Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.654594 4690 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.657471 4690 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.659259 4690 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 17 00:05:46 crc kubenswrapper[4690]: E0217 00:05:46.659521 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.686650 4690 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:55468->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.686733 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:55468->192.168.126.11:17697: read: connection reset by peer" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.687184 4690 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.687252 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.915988 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 11:09:43.027540291 +0000 UTC Feb 17 00:05:46 crc kubenswrapper[4690]: I0217 00:05:46.982215 4690 apiserver.go:52] "Watching apiserver" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.020260 4690 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.020674 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.021313 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.021337 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.021451 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.021510 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.021762 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.021847 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.022271 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.022335 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.022331 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.024308 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.024598 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.024607 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.024702 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.024785 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.024955 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.024998 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.025675 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.025772 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.069426 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.104048 4690 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.110760 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.122714 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.126455 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.127878 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764" exitCode=255 Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.127921 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764"} Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.141773 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.151733 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160215 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160288 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160324 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160384 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160417 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160447 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160477 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160506 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160554 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160623 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160664 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160691 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160749 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160790 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160827 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160856 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160885 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160980 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161018 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161047 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161077 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161107 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161139 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161169 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161221 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161259 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161328 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161390 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161428 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161461 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160621 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.160783 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161062 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161239 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161583 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161259 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161440 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161498 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161683 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161705 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161735 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161751 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161769 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161788 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161807 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161844 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161826 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161914 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161914 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161946 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161952 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161970 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162053 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162091 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162130 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162167 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162204 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162236 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162266 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162295 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162905 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162985 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163026 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163441 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163484 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163518 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163545 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163593 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163630 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163663 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163776 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163820 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163847 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163877 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163908 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163934 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163964 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163995 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164025 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164263 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164303 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164337 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164383 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164419 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164450 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164490 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164517 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164549 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164579 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164611 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164651 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164683 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164712 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164745 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164776 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164805 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164836 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164867 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164896 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164926 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162082 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164956 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162211 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163003 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163195 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.163510 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165042 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.162631 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164056 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.161530 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164761 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165076 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165116 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164849 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164921 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165066 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165245 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165279 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165384 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165560 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165626 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165621 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165670 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165672 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165832 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.165934 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:05:47.66589581 +0000 UTC m=+21.401608601 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.165970 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.166876 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.166986 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.168693 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.169176 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.169307 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.169566 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.169756 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.169927 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.169927 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.169937 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.170183 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.170474 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.170598 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.170808 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.171014 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.171516 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.164992 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.171751 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.171784 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.171827 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.171882 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.171935 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172023 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172076 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172112 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172149 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172252 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172262 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172307 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172551 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172615 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172604 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172733 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.172849 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.173425 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.173550 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.173631 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.173692 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.174284 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.174794 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.174929 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.174954 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.174982 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175043 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175077 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175651 4690 scope.go:117] "RemoveContainer" containerID="a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.173636 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175392 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175678 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175682 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175697 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175842 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175865 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175890 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175902 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175927 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175955 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176001 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176032 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176092 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176118 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176161 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176186 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176210 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176255 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176278 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176300 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176349 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176393 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176412 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176430 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176470 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176488 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176512 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176550 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176566 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176583 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176601 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176642 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176662 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176682 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176725 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177024 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177066 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177112 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177137 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177194 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177223 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177275 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177293 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177312 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177328 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177384 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177401 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177419 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177458 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177488 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177505 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177554 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177573 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177590 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177607 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177710 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177731 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177842 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177923 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177983 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178449 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178479 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178642 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178665 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178682 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175887 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176005 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176058 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.175923 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176160 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176172 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176186 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176225 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176308 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176378 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176455 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176784 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.176813 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177023 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177847 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.177976 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179191 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179222 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179239 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179329 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179349 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179504 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179524 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179544 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179581 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179602 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178054 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178311 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178418 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178427 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178868 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.178709 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179046 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179430 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179427 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179657 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179752 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179927 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.180086 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.180607 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.180647 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.180721 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.181214 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.181542 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.181542 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.181710 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182032 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182058 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182414 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.179817 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182481 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182505 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182523 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182535 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182568 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182588 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182588 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182667 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182694 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182711 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182750 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.183312 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.182815 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184073 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184432 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.183592 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.183766 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.183877 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184522 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184596 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184700 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184699 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.183585 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184790 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184797 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184931 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184937 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184961 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.184952 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185026 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185154 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185199 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185290 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185325 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185199 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185280 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185398 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185439 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185473 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185343 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185511 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185599 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185654 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185697 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185735 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185772 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185808 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185845 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185881 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185932 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.185969 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186004 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186058 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186092 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186127 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186203 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186233 4690 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186257 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186277 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186297 4690 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186316 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186336 4690 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186381 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186401 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186420 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186438 4690 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186457 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186478 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186496 4690 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186520 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186547 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186572 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186596 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186621 4690 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186649 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186673 4690 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.186695 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.186739 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.186781 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:47.686756863 +0000 UTC m=+21.422469704 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186699 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.186822 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:47.686799915 +0000 UTC m=+21.422512766 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186882 4690 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186953 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.186981 4690 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.187007 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.187029 4690 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.187164 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.187233 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.187422 4690 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.187740 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.187806 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.187862 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.187891 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.187910 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.188007 4690 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.188265 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.188761 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.189605 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.190244 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.188975 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.189790 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.190240 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.190279 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.190535 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.190547 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.191079 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.191422 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.191552 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.191527 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.191578 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.195806 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.196048 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.196171 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.196219 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.191010 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201081 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201127 4690 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201141 4690 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201154 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201193 4690 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201206 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201216 4690 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201228 4690 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201315 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201380 4690 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201416 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201427 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201455 4690 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201466 4690 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201476 4690 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201485 4690 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201495 4690 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201506 4690 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201522 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201534 4690 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201549 4690 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201559 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201568 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201577 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201587 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201599 4690 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201610 4690 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201620 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201629 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201641 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201649 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201659 4690 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201668 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201677 4690 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201686 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201695 4690 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201704 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201736 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201747 4690 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201756 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201766 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201775 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201783 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201793 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201820 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201830 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201840 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201848 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201860 4690 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201874 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201888 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201900 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201909 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201918 4690 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201927 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201936 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201944 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201953 4690 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201961 4690 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201971 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201980 4690 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.201993 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202001 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202010 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202018 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202027 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202035 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202044 4690 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202052 4690 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202060 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202069 4690 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202078 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202087 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202096 4690 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202106 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202115 4690 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202126 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202137 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202147 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202155 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202165 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202175 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202184 4690 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202193 4690 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202201 4690 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202210 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202218 4690 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202227 4690 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202238 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202251 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202262 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202274 4690 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202286 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202297 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202305 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202316 4690 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.202653 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.203080 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.203244 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.203299 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.203720 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.203891 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.203984 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.206318 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.206541 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.206836 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.207104 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.207475 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.207865 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.208003 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.208154 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.211081 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.211429 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.211461 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.211479 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.211553 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:47.711531787 +0000 UTC m=+21.447244548 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.211998 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.212109 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.212130 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.212143 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.212202 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:47.712179305 +0000 UTC m=+21.447892176 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.212196 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.212204 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.212687 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.215198 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.215424 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.215947 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.216036 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.220105 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.220165 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.220733 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.228161 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.229324 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.230696 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.230847 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.230997 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.231264 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.231434 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.231599 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.231625 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.232091 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.232189 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.232220 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.232586 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.232619 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.232864 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.232962 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.233296 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.233330 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.235126 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.235569 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.235643 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.236283 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.236774 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.244560 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.247198 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.253246 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.255329 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.255320 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.260879 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.267887 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.280089 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.289874 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.302902 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.302940 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.302974 4690 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.302984 4690 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.302993 4690 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303002 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303014 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303022 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303032 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303041 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303050 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303060 4690 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303071 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303080 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303089 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303099 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303109 4690 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303118 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303127 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303136 4690 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303146 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303155 4690 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303166 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303174 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303184 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303193 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303204 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303214 4690 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303224 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303232 4690 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303241 4690 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303250 4690 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303260 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303275 4690 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303285 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303294 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303303 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303313 4690 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303322 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303331 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303340 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303348 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303375 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303383 4690 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303391 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303402 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303410 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303419 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303427 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303436 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303444 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303454 4690 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303463 4690 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303473 4690 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303482 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303491 4690 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303499 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303508 4690 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303518 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303391 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303526 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303563 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303598 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303591 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303633 4690 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303668 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303686 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.303704 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.341583 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.358283 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 00:05:47 crc kubenswrapper[4690]: W0217 00:05:47.358638 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-9c9cebce5b35ee411ca4785161cfa8737e0b66c5c66bea68fcdef46c5db91c44 WatchSource:0}: Error finding container 9c9cebce5b35ee411ca4785161cfa8737e0b66c5c66bea68fcdef46c5db91c44: Status 404 returned error can't find the container with id 9c9cebce5b35ee411ca4785161cfa8737e0b66c5c66bea68fcdef46c5db91c44 Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.365603 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.494966 4690 csr.go:261] certificate signing request csr-7w5rj is approved, waiting to be issued Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.532983 4690 csr.go:257] certificate signing request csr-7w5rj is issued Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.617094 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.711474 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.711581 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.711609 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.711666 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:05:48.711622459 +0000 UTC m=+22.447335210 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.711702 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.711760 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.711768 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:48.711753443 +0000 UTC m=+22.447466184 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.711895 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.711991 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:48.711969029 +0000 UTC m=+22.447681780 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.712183 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.712208 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.712224 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.712278 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:48.712266797 +0000 UTC m=+22.447979618 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.813111 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.813256 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.813272 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.813283 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:47 crc kubenswrapper[4690]: E0217 00:05:47.813331 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:48.813315694 +0000 UTC m=+22.549028445 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.916317 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 17:27:25.062093726 +0000 UTC Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.996732 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-v5cn7"] Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.997041 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-v5cn7" Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.997776 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-lszwm"] Feb 17 00:05:47 crc kubenswrapper[4690]: I0217 00:05:47.997995 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.007150 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.007445 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.007592 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.007680 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.007761 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.007608 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.008503 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.008705 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.015021 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a9cb4ddb-30f5-476e-b815-a847fe68e60f-hosts-file\") pod \"node-resolver-v5cn7\" (UID: \"a9cb4ddb-30f5-476e-b815-a847fe68e60f\") " pod="openshift-dns/node-resolver-v5cn7" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.015070 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8868c633-9a82-4998-b1b4-06c6d76396ec-rootfs\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.015094 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8868c633-9a82-4998-b1b4-06c6d76396ec-proxy-tls\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.015129 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-527vz\" (UniqueName: \"kubernetes.io/projected/a9cb4ddb-30f5-476e-b815-a847fe68e60f-kube-api-access-527vz\") pod \"node-resolver-v5cn7\" (UID: \"a9cb4ddb-30f5-476e-b815-a847fe68e60f\") " pod="openshift-dns/node-resolver-v5cn7" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.015154 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8868c633-9a82-4998-b1b4-06c6d76396ec-mcd-auth-proxy-config\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.015173 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6gtd\" (UniqueName: \"kubernetes.io/projected/8868c633-9a82-4998-b1b4-06c6d76396ec-kube-api-access-s6gtd\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.019497 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.032992 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.040780 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.050747 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.058204 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.070439 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.080774 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.089034 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.107505 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.116019 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a9cb4ddb-30f5-476e-b815-a847fe68e60f-hosts-file\") pod \"node-resolver-v5cn7\" (UID: \"a9cb4ddb-30f5-476e-b815-a847fe68e60f\") " pod="openshift-dns/node-resolver-v5cn7" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.116055 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8868c633-9a82-4998-b1b4-06c6d76396ec-rootfs\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.116078 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8868c633-9a82-4998-b1b4-06c6d76396ec-proxy-tls\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.116112 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-527vz\" (UniqueName: \"kubernetes.io/projected/a9cb4ddb-30f5-476e-b815-a847fe68e60f-kube-api-access-527vz\") pod \"node-resolver-v5cn7\" (UID: \"a9cb4ddb-30f5-476e-b815-a847fe68e60f\") " pod="openshift-dns/node-resolver-v5cn7" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.116131 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8868c633-9a82-4998-b1b4-06c6d76396ec-mcd-auth-proxy-config\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.116144 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6gtd\" (UniqueName: \"kubernetes.io/projected/8868c633-9a82-4998-b1b4-06c6d76396ec-kube-api-access-s6gtd\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.116415 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a9cb4ddb-30f5-476e-b815-a847fe68e60f-hosts-file\") pod \"node-resolver-v5cn7\" (UID: \"a9cb4ddb-30f5-476e-b815-a847fe68e60f\") " pod="openshift-dns/node-resolver-v5cn7" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.117253 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8868c633-9a82-4998-b1b4-06c6d76396ec-mcd-auth-proxy-config\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.117349 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8868c633-9a82-4998-b1b4-06c6d76396ec-rootfs\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.123146 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8868c633-9a82-4998-b1b4-06c6d76396ec-proxy-tls\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.124474 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.130840 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"242833eabfe52b47da32434bd73e2fdd38b4440fc8779aa65e58c6d9a1393d06"} Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.132686 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081"} Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.132809 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390"} Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.132873 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9c9cebce5b35ee411ca4785161cfa8737e0b66c5c66bea68fcdef46c5db91c44"} Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.134804 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.136232 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.136894 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef"} Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.137558 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.139070 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d"} Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.139107 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9ff8a71a0f3505af4f1ef148aee4c88c711127e008af4691578a4b474eb4bc75"} Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.144912 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-527vz\" (UniqueName: \"kubernetes.io/projected/a9cb4ddb-30f5-476e-b815-a847fe68e60f-kube-api-access-527vz\") pod \"node-resolver-v5cn7\" (UID: \"a9cb4ddb-30f5-476e-b815-a847fe68e60f\") " pod="openshift-dns/node-resolver-v5cn7" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.148611 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6gtd\" (UniqueName: \"kubernetes.io/projected/8868c633-9a82-4998-b1b4-06c6d76396ec-kube-api-access-s6gtd\") pod \"machine-config-daemon-lszwm\" (UID: \"8868c633-9a82-4998-b1b4-06c6d76396ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.153465 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.167606 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.184024 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.211342 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.232617 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.250767 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.271022 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.304179 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.314980 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-v5cn7" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.321130 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:05:48 crc kubenswrapper[4690]: W0217 00:05:48.335894 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8868c633_9a82_4998_b1b4_06c6d76396ec.slice/crio-f4ecdddbd4e1a32ad8f6e348b8e89b9f3ec2493842627fcaf6fa105c239b76cd WatchSource:0}: Error finding container f4ecdddbd4e1a32ad8f6e348b8e89b9f3ec2493842627fcaf6fa105c239b76cd: Status 404 returned error can't find the container with id f4ecdddbd4e1a32ad8f6e348b8e89b9f3ec2493842627fcaf6fa105c239b76cd Feb 17 00:05:48 crc kubenswrapper[4690]: W0217 00:05:48.336483 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9cb4ddb_30f5_476e_b815_a847fe68e60f.slice/crio-5a272c3ee8bc5817a2f81b7aaf0594e54fe99eb2dcaafe77a346a9dc1caa4097 WatchSource:0}: Error finding container 5a272c3ee8bc5817a2f81b7aaf0594e54fe99eb2dcaafe77a346a9dc1caa4097: Status 404 returned error can't find the container with id 5a272c3ee8bc5817a2f81b7aaf0594e54fe99eb2dcaafe77a346a9dc1caa4097 Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.346110 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.377592 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.405744 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2frwb"] Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.406986 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-pw72j"] Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.407160 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.410062 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.410086 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.411067 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.411375 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.411439 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.412296 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.413909 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.415639 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-vb4v5"] Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.415938 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.416431 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.430498 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.430672 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.430831 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.431058 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.430985 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.431211 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.431259 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.431399 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.433785 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-netns\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.433839 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-var-lib-openvswitch\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.433871 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-node-log\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.433898 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5246f770-6fb0-4762-97c7-9473fcc26738-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.433921 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-multus-socket-dir-parent\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.433953 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-ovn-kubernetes\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.433983 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-etc-kubernetes\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.434034 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-slash\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.438761 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-systemd\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.438807 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-netd\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.438829 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-env-overrides\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.438852 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-var-lib-cni-multus\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.438877 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.438900 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-systemd-units\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.438920 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-etc-openvswitch\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.438940 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-ovn\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.438955 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-kubelet\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.438975 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-system-cni-dir\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439000 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-cnibin\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439027 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-var-lib-kubelet\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439052 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-run-multus-certs\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439120 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-run-k8s-cni-cncf-io\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439150 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqxq5\" (UniqueName: \"kubernetes.io/projected/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-kube-api-access-pqxq5\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439178 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5246f770-6fb0-4762-97c7-9473fcc26738-cni-binary-copy\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439217 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-os-release\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439245 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-bin\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439273 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-var-lib-cni-bin\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439301 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-multus-conf-dir\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439330 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-run-netns\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439379 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/797b14df-f7d0-419b-95f8-f02f25409e66-multus-daemon-config\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439401 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.439438 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-config\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.444558 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-script-lib\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.444596 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfv9x\" (UniqueName: \"kubernetes.io/projected/797b14df-f7d0-419b-95f8-f02f25409e66-kube-api-access-hfv9x\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.445714 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-multus-cni-dir\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.445755 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/797b14df-f7d0-419b-95f8-f02f25409e66-cni-binary-copy\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.445785 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-hostroot\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.445810 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-openvswitch\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.445830 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-system-cni-dir\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.445855 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-log-socket\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.445887 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2lxb\" (UniqueName: \"kubernetes.io/projected/5246f770-6fb0-4762-97c7-9473fcc26738-kube-api-access-f2lxb\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.445917 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovn-node-metrics-cert\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.445931 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-os-release\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.445967 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-cnibin\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.461013 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.484791 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.497258 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.518427 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.533592 4690 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-17 00:00:47 +0000 UTC, rotation deadline is 2026-10-30 23:30:16.9907673 +0000 UTC Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.533644 4690 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6143h24m28.457126101s for next certificate rotation Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.540361 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.546365 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-bin\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.546562 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-var-lib-cni-bin\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.546644 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-var-lib-cni-bin\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.546504 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-bin\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.546661 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-multus-conf-dir\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.546895 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.547015 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-config\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.547127 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-script-lib\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.547229 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-run-netns\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.547335 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/797b14df-f7d0-419b-95f8-f02f25409e66-multus-daemon-config\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.547498 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfv9x\" (UniqueName: \"kubernetes.io/projected/797b14df-f7d0-419b-95f8-f02f25409e66-kube-api-access-hfv9x\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548037 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-multus-cni-dir\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548140 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/797b14df-f7d0-419b-95f8-f02f25409e66-cni-binary-copy\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548248 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-openvswitch\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548339 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-system-cni-dir\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548490 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-hostroot\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548592 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-log-socket\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548686 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2lxb\" (UniqueName: \"kubernetes.io/projected/5246f770-6fb0-4762-97c7-9473fcc26738-kube-api-access-f2lxb\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548781 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovn-node-metrics-cert\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548863 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-os-release\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548968 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-cnibin\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549061 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5246f770-6fb0-4762-97c7-9473fcc26738-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549153 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-multus-socket-dir-parent\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549242 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-netns\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549329 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-var-lib-openvswitch\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549446 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-node-log\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549542 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-ovn-kubernetes\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549654 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-etc-kubernetes\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549749 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-env-overrides\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549841 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-var-lib-cni-multus\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549938 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-slash\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550029 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-systemd\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550119 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-netd\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550208 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-systemd-units\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550304 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-etc-openvswitch\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550387 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-ovn\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550505 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-node-log\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550510 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550587 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-var-lib-kubelet\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550606 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-run-multus-certs\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550649 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-kubelet\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550666 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-system-cni-dir\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550685 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-cnibin\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550706 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqxq5\" (UniqueName: \"kubernetes.io/projected/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-kube-api-access-pqxq5\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550722 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5246f770-6fb0-4762-97c7-9473fcc26738-cni-binary-copy\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550731 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5246f770-6fb0-4762-97c7-9473fcc26738-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550738 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-run-k8s-cni-cncf-io\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550778 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-os-release\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550792 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-multus-socket-dir-parent\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549254 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/797b14df-f7d0-419b-95f8-f02f25409e66-cni-binary-copy\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550846 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-var-lib-openvswitch\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550864 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-ovn-kubernetes\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.546819 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-multus-conf-dir\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.549559 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-multus-cni-dir\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550122 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-cnibin\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.547388 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-run-netns\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550956 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-kubelet\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550966 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-systemd\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550975 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-etc-kubernetes\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550085 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-os-release\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550996 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-netd\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551001 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-system-cni-dir\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551017 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-cnibin\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551031 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-etc-openvswitch\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551059 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-systemd-units\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551079 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-var-lib-kubelet\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551288 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-var-lib-cni-multus\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.546975 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551442 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-env-overrides\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.547883 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-script-lib\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.547759 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-config\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.548207 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/797b14df-f7d0-419b-95f8-f02f25409e66-multus-daemon-config\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.550476 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-slash\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551511 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-run-k8s-cni-cncf-io\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551542 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-os-release\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551561 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-ovn\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551572 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-log-socket\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551568 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-openvswitch\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551588 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-netns\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551596 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-hostroot\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551612 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-host-run-multus-certs\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551694 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/797b14df-f7d0-419b-95f8-f02f25409e66-system-cni-dir\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.551801 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5246f770-6fb0-4762-97c7-9473fcc26738-cni-binary-copy\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.552832 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5246f770-6fb0-4762-97c7-9473fcc26738-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.554504 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovn-node-metrics-cert\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.557704 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.566705 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfv9x\" (UniqueName: \"kubernetes.io/projected/797b14df-f7d0-419b-95f8-f02f25409e66-kube-api-access-hfv9x\") pod \"multus-vb4v5\" (UID: \"797b14df-f7d0-419b-95f8-f02f25409e66\") " pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.568702 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2lxb\" (UniqueName: \"kubernetes.io/projected/5246f770-6fb0-4762-97c7-9473fcc26738-kube-api-access-f2lxb\") pod \"multus-additional-cni-plugins-pw72j\" (UID: \"5246f770-6fb0-4762-97c7-9473fcc26738\") " pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.570753 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.573633 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqxq5\" (UniqueName: \"kubernetes.io/projected/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-kube-api-access-pqxq5\") pod \"ovnkube-node-2frwb\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.591965 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.611795 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.630661 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.650332 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.663552 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.675821 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.686505 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.698039 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.709995 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:48Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.731309 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:48 crc kubenswrapper[4690]: W0217 00:05:48.743953 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6eeb20f_1ee2_4d96_a53b_da30d60213f3.slice/crio-8e53d779a45111fcc088635b7ab3cbba7b3db883020c91984e6348af43fab072 WatchSource:0}: Error finding container 8e53d779a45111fcc088635b7ab3cbba7b3db883020c91984e6348af43fab072: Status 404 returned error can't find the container with id 8e53d779a45111fcc088635b7ab3cbba7b3db883020c91984e6348af43fab072 Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.752016 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.752148 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.752183 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.752213 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.752344 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.752362 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.752376 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.752455 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:50.752442391 +0000 UTC m=+24.488155142 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.752665 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.752769 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:05:50.75273914 +0000 UTC m=+24.488451891 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.752879 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:50.752850723 +0000 UTC m=+24.488563504 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.753023 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.753214 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:50.753191802 +0000 UTC m=+24.488904603 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.755726 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vb4v5" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.762176 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pw72j" Feb 17 00:05:48 crc kubenswrapper[4690]: W0217 00:05:48.780235 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod797b14df_f7d0_419b_95f8_f02f25409e66.slice/crio-50361e1c6850d4e7ed173a4ef9ca4e89dee15853d4c599ded92bc89010414225 WatchSource:0}: Error finding container 50361e1c6850d4e7ed173a4ef9ca4e89dee15853d4c599ded92bc89010414225: Status 404 returned error can't find the container with id 50361e1c6850d4e7ed173a4ef9ca4e89dee15853d4c599ded92bc89010414225 Feb 17 00:05:48 crc kubenswrapper[4690]: W0217 00:05:48.784092 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5246f770_6fb0_4762_97c7_9473fcc26738.slice/crio-ad95620fc1f7866d2e232b9276b5ba24a777e7419f6fd20a6d586bfbabab542b WatchSource:0}: Error finding container ad95620fc1f7866d2e232b9276b5ba24a777e7419f6fd20a6d586bfbabab542b: Status 404 returned error can't find the container with id ad95620fc1f7866d2e232b9276b5ba24a777e7419f6fd20a6d586bfbabab542b Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.853108 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.853247 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.853270 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.853282 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.853322 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:50.853310233 +0000 UTC m=+24.589022985 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.917258 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 08:15:17.899143473 +0000 UTC Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.977393 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.977458 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.977591 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.977636 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.977715 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:05:48 crc kubenswrapper[4690]: E0217 00:05:48.977767 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.981449 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.982190 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.983268 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.983901 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.984817 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.985323 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.986210 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.987353 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.987985 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.988981 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.989513 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.991160 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.992533 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.993084 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.994109 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.994787 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.996005 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.996559 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.997121 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.998113 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 17 00:05:48 crc kubenswrapper[4690]: I0217 00:05:48.998776 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.000115 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.000702 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.001566 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.002272 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.003092 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.004953 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.005507 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.006246 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.006781 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.007334 4690 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.008261 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.009926 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.010506 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.011374 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.012939 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.013591 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.014549 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.015218 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.016280 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.016793 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.017757 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.018405 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.019492 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.020010 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.021028 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.021673 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.022794 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.023286 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.024159 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.024732 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.025673 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.026638 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.027096 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.143557 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9" exitCode=0 Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.143633 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.143679 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"8e53d779a45111fcc088635b7ab3cbba7b3db883020c91984e6348af43fab072"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.146500 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" event={"ID":"5246f770-6fb0-4762-97c7-9473fcc26738","Type":"ContainerStarted","Data":"1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.146557 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" event={"ID":"5246f770-6fb0-4762-97c7-9473fcc26738","Type":"ContainerStarted","Data":"ad95620fc1f7866d2e232b9276b5ba24a777e7419f6fd20a6d586bfbabab542b"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.152979 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vb4v5" event={"ID":"797b14df-f7d0-419b-95f8-f02f25409e66","Type":"ContainerStarted","Data":"727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.153031 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vb4v5" event={"ID":"797b14df-f7d0-419b-95f8-f02f25409e66","Type":"ContainerStarted","Data":"50361e1c6850d4e7ed173a4ef9ca4e89dee15853d4c599ded92bc89010414225"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.154796 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-v5cn7" event={"ID":"a9cb4ddb-30f5-476e-b815-a847fe68e60f","Type":"ContainerStarted","Data":"6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.154828 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-v5cn7" event={"ID":"a9cb4ddb-30f5-476e-b815-a847fe68e60f","Type":"ContainerStarted","Data":"5a272c3ee8bc5817a2f81b7aaf0594e54fe99eb2dcaafe77a346a9dc1caa4097"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.162291 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerStarted","Data":"e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.162606 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerStarted","Data":"e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.162729 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerStarted","Data":"f4ecdddbd4e1a32ad8f6e348b8e89b9f3ec2493842627fcaf6fa105c239b76cd"} Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.163425 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.178937 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.197278 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.212998 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.228609 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.250698 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.294295 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.344899 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.389633 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.406316 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.419145 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.437661 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.448238 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.464653 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.483915 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.497324 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.529912 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.540066 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.553143 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.571180 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.582471 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.596296 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.607855 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.621699 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.918200 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 22:56:38.398096335 +0000 UTC Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.953183 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-wpnhj"] Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.953893 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.955774 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.955825 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.956410 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.957292 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.970920 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:49 crc kubenswrapper[4690]: I0217 00:05:49.987620 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.001016 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:49Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.018654 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.033881 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.050164 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.066496 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.072579 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpndk\" (UniqueName: \"kubernetes.io/projected/ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30-kube-api-access-lpndk\") pod \"node-ca-wpnhj\" (UID: \"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\") " pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.072632 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30-serviceca\") pod \"node-ca-wpnhj\" (UID: \"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\") " pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.072913 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30-host\") pod \"node-ca-wpnhj\" (UID: \"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\") " pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.080106 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.094833 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.115188 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.130459 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.148433 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.162661 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.172319 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8"} Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.172395 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b"} Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.172409 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4"} Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.172421 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf"} Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.172432 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869"} Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.173369 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30-host\") pod \"node-ca-wpnhj\" (UID: \"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\") " pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.173428 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30-serviceca\") pod \"node-ca-wpnhj\" (UID: \"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\") " pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.173448 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpndk\" (UniqueName: \"kubernetes.io/projected/ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30-kube-api-access-lpndk\") pod \"node-ca-wpnhj\" (UID: \"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\") " pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.173644 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30-host\") pod \"node-ca-wpnhj\" (UID: \"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\") " pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.176254 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30-serviceca\") pod \"node-ca-wpnhj\" (UID: \"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\") " pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.177865 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4"} Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.189594 4690 generic.go:334] "Generic (PLEG): container finished" podID="5246f770-6fb0-4762-97c7-9473fcc26738" containerID="1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171" exitCode=0 Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.189665 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" event={"ID":"5246f770-6fb0-4762-97c7-9473fcc26738","Type":"ContainerDied","Data":"1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171"} Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.201327 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpndk\" (UniqueName: \"kubernetes.io/projected/ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30-kube-api-access-lpndk\") pod \"node-ca-wpnhj\" (UID: \"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\") " pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.203215 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.221129 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.246105 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.261274 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.273768 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.286493 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.298310 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.314574 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.330048 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.335349 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wpnhj" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.349177 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: W0217 00:05:50.371192 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea8d7ce4_2bc3_4db6_8a8e_467d9a5bac30.slice/crio-1ff1c573d40f2cdb5c323c02b8cbfee5804b63fafc4789c7d8edf35bd360c77b WatchSource:0}: Error finding container 1ff1c573d40f2cdb5c323c02b8cbfee5804b63fafc4789c7d8edf35bd360c77b: Status 404 returned error can't find the container with id 1ff1c573d40f2cdb5c323c02b8cbfee5804b63fafc4789c7d8edf35bd360c77b Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.384769 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.398353 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.414275 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.427897 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.441446 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.459230 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.474454 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.487261 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.498543 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.512527 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.530236 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.543468 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.568193 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.611944 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.650722 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.689289 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.781352 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.781514 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.781559 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.781626 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.781713 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:05:54.781667248 +0000 UTC m=+28.517380039 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.781739 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.781771 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.781778 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.781786 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.781837 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:54.781822362 +0000 UTC m=+28.517535113 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.781855 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:54.781848353 +0000 UTC m=+28.517561104 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.781898 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.782005 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:54.781976297 +0000 UTC m=+28.517689048 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.882666 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.882821 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.882847 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.882859 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.882922 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 00:05:54.88290296 +0000 UTC m=+28.618615711 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.920123 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 00:43:56.25065106 +0000 UTC Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.975898 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.976052 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.976497 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.976574 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:05:50 crc kubenswrapper[4690]: I0217 00:05:50.976629 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:50 crc kubenswrapper[4690]: E0217 00:05:50.976680 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.218269 4690 generic.go:334] "Generic (PLEG): container finished" podID="5246f770-6fb0-4762-97c7-9473fcc26738" containerID="21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6" exitCode=0 Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.218359 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" event={"ID":"5246f770-6fb0-4762-97c7-9473fcc26738","Type":"ContainerDied","Data":"21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6"} Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.222748 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616"} Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.225314 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wpnhj" event={"ID":"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30","Type":"ContainerStarted","Data":"a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c"} Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.225369 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wpnhj" event={"ID":"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30","Type":"ContainerStarted","Data":"1ff1c573d40f2cdb5c323c02b8cbfee5804b63fafc4789c7d8edf35bd360c77b"} Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.264363 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.286031 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.307184 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.327737 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.338332 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.349599 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.372724 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.388223 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.400406 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.418099 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.431295 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.442337 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.455578 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.469081 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.480840 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.497323 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.510852 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.523820 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.538606 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.549071 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.564461 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.585466 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.609906 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.655069 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.693871 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.727618 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:51Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:51 crc kubenswrapper[4690]: I0217 00:05:51.920298 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 18:34:01.431778308 +0000 UTC Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.232259 4690 generic.go:334] "Generic (PLEG): container finished" podID="5246f770-6fb0-4762-97c7-9473fcc26738" containerID="9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e" exitCode=0 Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.232304 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" event={"ID":"5246f770-6fb0-4762-97c7-9473fcc26738","Type":"ContainerDied","Data":"9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e"} Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.246514 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.262260 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.277035 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.291669 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.304439 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.316552 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.325328 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.336194 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.360713 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.380524 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.395800 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.410828 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.422915 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.630578 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.635650 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.639994 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.646984 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.658487 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.671313 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.685102 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.699018 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.710773 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.724117 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.733581 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.743819 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.766535 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.781678 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.795449 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.805138 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.830676 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.870488 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.921223 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 11:01:59.222906613 +0000 UTC Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.931329 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.964141 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.975923 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:52 crc kubenswrapper[4690]: E0217 00:05:52.976047 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.976371 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:52 crc kubenswrapper[4690]: E0217 00:05:52.976448 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.976470 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:52 crc kubenswrapper[4690]: E0217 00:05:52.976516 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:05:52 crc kubenswrapper[4690]: I0217 00:05:52.989692 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.028927 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.059725 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.061617 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.061654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.061665 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.061797 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.069776 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.120018 4690 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.120324 4690 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.121449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.121510 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.121525 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.121545 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.121558 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: E0217 00:05:53.132948 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.136344 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.136398 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.136412 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.136433 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.136446 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.146837 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: E0217 00:05:53.147669 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.151840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.151897 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.151911 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.151937 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.151950 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: E0217 00:05:53.169039 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.173615 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.173662 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.173674 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.173690 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.173700 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: E0217 00:05:53.185864 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.186827 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.189173 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.189212 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.189226 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.189248 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.189267 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: E0217 00:05:53.203026 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: E0217 00:05:53.203143 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.204579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.204624 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.204639 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.204657 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.204671 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.227059 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.240085 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.242351 4690 generic.go:334] "Generic (PLEG): container finished" podID="5246f770-6fb0-4762-97c7-9473fcc26738" containerID="59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47" exitCode=0 Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.242417 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" event={"ID":"5246f770-6fb0-4762-97c7-9473fcc26738","Type":"ContainerDied","Data":"59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.272885 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.307455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.307504 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.307516 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.307538 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.307551 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.311005 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.349858 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.387685 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.409334 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.409398 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.409410 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.409427 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.409439 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.428612 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.465388 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.511334 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.513458 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.513495 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.513513 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.513538 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.513558 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.557792 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.590910 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.615744 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.615831 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.615859 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.615892 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.615918 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.630355 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.668121 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.707242 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.718520 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.718559 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.718569 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.718585 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.718623 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.747894 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.786805 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.820610 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.820865 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.820949 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.821049 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.821113 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.833716 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.872090 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.914870 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.921688 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 13:14:06.723683416 +0000 UTC Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.924046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.924093 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.924112 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.924137 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.924158 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:53Z","lastTransitionTime":"2026-02-17T00:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:53 crc kubenswrapper[4690]: I0217 00:05:53.947739 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:53Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.026946 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.027005 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.027027 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.027053 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.027070 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:54Z","lastTransitionTime":"2026-02-17T00:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.130314 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.130359 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.130388 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.130408 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.130420 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:54Z","lastTransitionTime":"2026-02-17T00:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.232879 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.232923 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.232933 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.232951 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.232962 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:54Z","lastTransitionTime":"2026-02-17T00:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.250001 4690 generic.go:334] "Generic (PLEG): container finished" podID="5246f770-6fb0-4762-97c7-9473fcc26738" containerID="769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a" exitCode=0 Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.250035 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" event={"ID":"5246f770-6fb0-4762-97c7-9473fcc26738","Type":"ContainerDied","Data":"769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.274564 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.293286 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.309603 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.328161 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.335695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.335757 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.335776 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.335802 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.335820 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:54Z","lastTransitionTime":"2026-02-17T00:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.343960 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.359091 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.371415 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.391406 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.407282 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.424758 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.438202 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.438285 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.438319 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.438381 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.438398 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:54Z","lastTransitionTime":"2026-02-17T00:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.438479 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.450481 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.467907 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.508316 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.540934 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.540978 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.540993 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.541010 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.541023 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:54Z","lastTransitionTime":"2026-02-17T00:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.643338 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.643433 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.643447 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.643465 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.643480 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:54Z","lastTransitionTime":"2026-02-17T00:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.747792 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.747825 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.747833 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.747846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.747855 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:54Z","lastTransitionTime":"2026-02-17T00:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.850200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.850247 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.850259 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.850277 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.850290 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:54Z","lastTransitionTime":"2026-02-17T00:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.853097 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.853251 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:06:02.853230499 +0000 UTC m=+36.588943260 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.853299 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.853353 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.853419 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.853500 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.853519 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.853569 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:02.853560109 +0000 UTC m=+36.589272870 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.853589 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.853596 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:02.853577369 +0000 UTC m=+36.589290160 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.853613 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.853626 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.853685 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:02.853663882 +0000 UTC m=+36.589376723 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.922423 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 23:21:16.946988161 +0000 UTC Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.952505 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.952563 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.952579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.952601 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.952616 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:54Z","lastTransitionTime":"2026-02-17T00:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.954164 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.954403 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.954438 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.954456 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.954581 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:02.954511013 +0000 UTC m=+36.690223814 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.979123 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.979286 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.979986 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.980126 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:05:54 crc kubenswrapper[4690]: I0217 00:05:54.980206 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:54 crc kubenswrapper[4690]: E0217 00:05:54.980283 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.055700 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.055750 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.055770 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.055792 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.055829 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:55Z","lastTransitionTime":"2026-02-17T00:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.166619 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.166727 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.166747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.166778 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.166799 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:55Z","lastTransitionTime":"2026-02-17T00:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.260945 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.261628 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.261825 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.267031 4690 generic.go:334] "Generic (PLEG): container finished" podID="5246f770-6fb0-4762-97c7-9473fcc26738" containerID="04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3" exitCode=0 Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.267084 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" event={"ID":"5246f770-6fb0-4762-97c7-9473fcc26738","Type":"ContainerDied","Data":"04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.269145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.269185 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.269203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.269227 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.269245 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:55Z","lastTransitionTime":"2026-02-17T00:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.290351 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.309672 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.310116 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.312771 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.332597 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.359320 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.372818 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.372876 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.372892 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.372959 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.372973 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:55Z","lastTransitionTime":"2026-02-17T00:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.378628 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.396028 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.411685 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.456326 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.476207 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.477562 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.477586 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.477595 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.477608 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.477617 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:55Z","lastTransitionTime":"2026-02-17T00:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.487989 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.502494 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.519107 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.536354 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.553199 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.570604 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.580252 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.580296 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.580308 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.580330 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.580344 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:55Z","lastTransitionTime":"2026-02-17T00:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.584404 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.598923 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.612961 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.626272 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.637723 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.655302 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.673044 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.682312 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.682384 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.682409 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.682429 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.682443 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:55Z","lastTransitionTime":"2026-02-17T00:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.688658 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.703193 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.713711 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.728810 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.781610 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.785397 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.785444 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.785458 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.785482 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.785499 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:55Z","lastTransitionTime":"2026-02-17T00:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.804754 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:55Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.888885 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.888938 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.888949 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.888969 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.888981 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:55Z","lastTransitionTime":"2026-02-17T00:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.922721 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 14:32:41.166833713 +0000 UTC Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.991602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.991655 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.991678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.991704 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:55 crc kubenswrapper[4690]: I0217 00:05:55.991723 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:55Z","lastTransitionTime":"2026-02-17T00:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.094252 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.094293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.094306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.094323 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.094334 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:56Z","lastTransitionTime":"2026-02-17T00:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.197424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.197476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.197494 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.197519 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.197537 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:56Z","lastTransitionTime":"2026-02-17T00:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.274739 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" event={"ID":"5246f770-6fb0-4762-97c7-9473fcc26738","Type":"ContainerStarted","Data":"21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.274815 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.297687 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.299696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.299747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.299760 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.299779 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.299791 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:56Z","lastTransitionTime":"2026-02-17T00:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.318740 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.335212 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.352267 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.366900 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.385117 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.403782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.403816 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.403826 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.403839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.403847 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:56Z","lastTransitionTime":"2026-02-17T00:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.405516 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.427891 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.447978 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.467231 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.479991 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.498502 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.506897 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.506945 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.506962 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.506985 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.507002 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:56Z","lastTransitionTime":"2026-02-17T00:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.514230 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.534708 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.610954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.611000 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.611014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.611034 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.611048 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:56Z","lastTransitionTime":"2026-02-17T00:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.714168 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.714231 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.714293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.714319 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.714337 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:56Z","lastTransitionTime":"2026-02-17T00:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.751418 4690 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.817087 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.817122 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.817134 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.817151 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.817163 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:56Z","lastTransitionTime":"2026-02-17T00:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.920473 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.920507 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.920519 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.920560 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.920573 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:56Z","lastTransitionTime":"2026-02-17T00:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.923034 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 10:35:48.7202918 +0000 UTC Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.975623 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:56 crc kubenswrapper[4690]: E0217 00:05:56.975804 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.976556 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:56 crc kubenswrapper[4690]: E0217 00:05:56.976682 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.977299 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:56 crc kubenswrapper[4690]: E0217 00:05:56.977452 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:05:56 crc kubenswrapper[4690]: I0217 00:05:56.994594 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:56Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.011553 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.023458 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.023503 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.023515 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.023532 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.023544 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:57Z","lastTransitionTime":"2026-02-17T00:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.031285 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.056199 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.070588 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.084943 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.095866 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.107648 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.125345 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.125390 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.125398 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.125411 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.125420 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:57Z","lastTransitionTime":"2026-02-17T00:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.129329 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.144069 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.157529 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.183275 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.201230 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.213575 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.228325 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.228377 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.228387 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.228403 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.228412 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:57Z","lastTransitionTime":"2026-02-17T00:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.277693 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.331300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.331382 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.331400 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.331418 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.331430 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:57Z","lastTransitionTime":"2026-02-17T00:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.438092 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.438143 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.438155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.438177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.438190 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:57Z","lastTransitionTime":"2026-02-17T00:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.541200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.541234 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.541243 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.541257 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.541267 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:57Z","lastTransitionTime":"2026-02-17T00:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.644474 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.644549 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.644569 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.644597 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.644618 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:57Z","lastTransitionTime":"2026-02-17T00:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.747579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.747665 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.747683 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.747731 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.747756 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:57Z","lastTransitionTime":"2026-02-17T00:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.850426 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.850528 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.850556 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.850591 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.850621 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:57Z","lastTransitionTime":"2026-02-17T00:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.924415 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 14:17:57.35731163 +0000 UTC Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.953668 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.953737 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.953759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.953787 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:57 crc kubenswrapper[4690]: I0217 00:05:57.953810 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:57Z","lastTransitionTime":"2026-02-17T00:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.055875 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.055920 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.055931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.055950 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.055962 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:58Z","lastTransitionTime":"2026-02-17T00:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.158755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.158804 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.158815 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.158833 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.158843 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:58Z","lastTransitionTime":"2026-02-17T00:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.261448 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.261483 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.261493 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.261512 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.261523 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:58Z","lastTransitionTime":"2026-02-17T00:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.284749 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/0.log" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.289014 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3" exitCode=1 Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.289069 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.289878 4690 scope.go:117] "RemoveContainer" containerID="1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.308806 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.330753 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.351503 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.364864 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.364914 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.364931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.364958 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.364975 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:58Z","lastTransitionTime":"2026-02-17T00:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.369613 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.392026 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.411070 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.432207 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:57Z\\\",\\\"message\\\":\\\"0:05:57.630446 6010 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:57.631158 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:57.631182 6010 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:57.633321 6010 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 00:05:57.633334 6010 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 00:05:57.633385 6010 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0217 00:05:57.633396 6010 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 00:05:57.633413 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:57.633416 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:57.633409 6010 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 00:05:57.633435 6010 factory.go:656] Stopping watch factory\\\\nI0217 00:05:57.633445 6010 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:57.633458 6010 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:57.633465 6010 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:57.633480 6010 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0217 00:05:57.633479 6010 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.446668 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.459822 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.468787 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.468826 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.468839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.468859 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.468878 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:58Z","lastTransitionTime":"2026-02-17T00:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.479817 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.492931 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.504864 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.515888 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.526246 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:58Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.571044 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.571076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.571083 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.571097 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.571106 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:58Z","lastTransitionTime":"2026-02-17T00:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.673300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.673346 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.673393 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.673414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.673427 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:58Z","lastTransitionTime":"2026-02-17T00:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.806475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.806522 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.806534 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.806551 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.806563 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:58Z","lastTransitionTime":"2026-02-17T00:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.908666 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.908699 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.908707 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.908722 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.908730 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:58Z","lastTransitionTime":"2026-02-17T00:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.924945 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 15:55:54.662745895 +0000 UTC Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.975607 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.975648 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:05:58 crc kubenswrapper[4690]: E0217 00:05:58.975823 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:05:58 crc kubenswrapper[4690]: I0217 00:05:58.975843 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:05:58 crc kubenswrapper[4690]: E0217 00:05:58.975972 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:05:58 crc kubenswrapper[4690]: E0217 00:05:58.976080 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.011257 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.011305 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.011318 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.011333 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.011345 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:59Z","lastTransitionTime":"2026-02-17T00:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.113957 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.114402 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.114600 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.114788 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.114966 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:59Z","lastTransitionTime":"2026-02-17T00:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.216963 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.217000 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.217010 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.217027 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.217044 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:59Z","lastTransitionTime":"2026-02-17T00:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.295879 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/0.log" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.298936 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f"} Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.299078 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.319186 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.319960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.320005 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.320018 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.320036 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.320048 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:59Z","lastTransitionTime":"2026-02-17T00:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.338431 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.356048 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.369979 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.380571 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.391000 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.407818 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:57Z\\\",\\\"message\\\":\\\"0:05:57.630446 6010 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:57.631158 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:57.631182 6010 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:57.633321 6010 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 00:05:57.633334 6010 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 00:05:57.633385 6010 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0217 00:05:57.633396 6010 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 00:05:57.633413 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:57.633416 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:57.633409 6010 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 00:05:57.633435 6010 factory.go:656] Stopping watch factory\\\\nI0217 00:05:57.633445 6010 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:57.633458 6010 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:57.633465 6010 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:57.633480 6010 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0217 00:05:57.633479 6010 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.423024 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.423075 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.423087 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.423106 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.423118 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:59Z","lastTransitionTime":"2026-02-17T00:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.429295 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.443669 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.456093 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.469697 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.484802 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.498019 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.513964 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:05:59Z is after 2025-08-24T17:21:41Z" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.526172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.526310 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.526327 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.526345 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.526379 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:59Z","lastTransitionTime":"2026-02-17T00:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.629021 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.629076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.629089 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.629107 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.629120 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:59Z","lastTransitionTime":"2026-02-17T00:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.731927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.731982 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.731997 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.732014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.732023 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:59Z","lastTransitionTime":"2026-02-17T00:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.834779 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.834833 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.834845 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.834862 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.834875 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:59Z","lastTransitionTime":"2026-02-17T00:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.925674 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 14:22:24.854608368 +0000 UTC Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.937721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.937775 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.937792 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.937822 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:05:59 crc kubenswrapper[4690]: I0217 00:05:59.937843 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:05:59Z","lastTransitionTime":"2026-02-17T00:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.040275 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.040341 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.040428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.040462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.040481 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:00Z","lastTransitionTime":"2026-02-17T00:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.143293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.143387 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.143405 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.143430 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.143447 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:00Z","lastTransitionTime":"2026-02-17T00:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.245972 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.246045 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.246069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.246098 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.246118 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:00Z","lastTransitionTime":"2026-02-17T00:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.304693 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/1.log" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.305259 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/0.log" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.308179 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f" exitCode=1 Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.308230 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.308273 4690 scope.go:117] "RemoveContainer" containerID="1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.309660 4690 scope.go:117] "RemoveContainer" containerID="520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f" Feb 17 00:06:00 crc kubenswrapper[4690]: E0217 00:06:00.309974 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.329570 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.347955 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.348152 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.348208 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.348225 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.348253 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.348270 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:00Z","lastTransitionTime":"2026-02-17T00:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.363066 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.380538 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.409719 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:57Z\\\",\\\"message\\\":\\\"0:05:57.630446 6010 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:57.631158 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:57.631182 6010 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:57.633321 6010 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 00:05:57.633334 6010 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 00:05:57.633385 6010 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0217 00:05:57.633396 6010 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 00:05:57.633413 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:57.633416 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:57.633409 6010 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 00:05:57.633435 6010 factory.go:656] Stopping watch factory\\\\nI0217 00:05:57.633445 6010 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:57.633458 6010 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:57.633465 6010 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:57.633480 6010 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0217 00:05:57.633479 6010 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:59Z\\\",\\\"message\\\":\\\"190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 00:05:59.305970 6137 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 00:05:59.305982 6137 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:59.305986 6137 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:59.305984 6137 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:59.305987 6137 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:59.306007 6137 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:59.306014 6137 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:59.306022 6137 factory.go:656] Stopping watch factory\\\\nI0217 00:05:59.306023 6137 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 00:05:59.306032 6137 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:59.306050 6137 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0217 00:05:59.306958 6137 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0217 00:05:59.307056 6137 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0217 00:05:59.307087 6137 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:59.307116 6137 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 00:05:59.307204 6137 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.433313 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.447679 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.450394 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.450432 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.450441 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.450456 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.450465 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:00Z","lastTransitionTime":"2026-02-17T00:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.461209 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.474118 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.485182 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.496505 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.511201 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.522679 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.537991 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:00Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.552539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.552567 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.552576 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.552590 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.552599 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:00Z","lastTransitionTime":"2026-02-17T00:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.655483 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.655547 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.655585 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.655658 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.655686 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:00Z","lastTransitionTime":"2026-02-17T00:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.759882 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.759931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.759943 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.759962 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.759974 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:00Z","lastTransitionTime":"2026-02-17T00:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.862575 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.862639 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.862656 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.862683 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.862703 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:00Z","lastTransitionTime":"2026-02-17T00:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.927316 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 07:19:40.831680766 +0000 UTC Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.965792 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.965820 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.965828 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.965840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.965850 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:00Z","lastTransitionTime":"2026-02-17T00:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.976699 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.976807 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:00 crc kubenswrapper[4690]: E0217 00:06:00.976891 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:00 crc kubenswrapper[4690]: E0217 00:06:00.976961 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:00 crc kubenswrapper[4690]: I0217 00:06:00.977065 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:00 crc kubenswrapper[4690]: E0217 00:06:00.977134 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.068214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.068257 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.068268 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.068284 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.068294 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:01Z","lastTransitionTime":"2026-02-17T00:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.170738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.170788 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.170803 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.170823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.170834 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:01Z","lastTransitionTime":"2026-02-17T00:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.273767 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.273816 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.273828 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.273847 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.273859 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:01Z","lastTransitionTime":"2026-02-17T00:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.313418 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/1.log" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.376709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.376764 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.376775 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.376795 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.376808 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:01Z","lastTransitionTime":"2026-02-17T00:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.395105 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.410921 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.422942 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.436696 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.459551 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:57Z\\\",\\\"message\\\":\\\"0:05:57.630446 6010 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:57.631158 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:57.631182 6010 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:57.633321 6010 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 00:05:57.633334 6010 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 00:05:57.633385 6010 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0217 00:05:57.633396 6010 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 00:05:57.633413 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:57.633416 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:57.633409 6010 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 00:05:57.633435 6010 factory.go:656] Stopping watch factory\\\\nI0217 00:05:57.633445 6010 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:57.633458 6010 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:57.633465 6010 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:57.633480 6010 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0217 00:05:57.633479 6010 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:59Z\\\",\\\"message\\\":\\\"190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 00:05:59.305970 6137 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 00:05:59.305982 6137 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:59.305986 6137 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:59.305984 6137 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:59.305987 6137 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:59.306007 6137 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:59.306014 6137 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:59.306022 6137 factory.go:656] Stopping watch factory\\\\nI0217 00:05:59.306023 6137 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 00:05:59.306032 6137 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:59.306050 6137 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0217 00:05:59.306958 6137 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0217 00:05:59.307056 6137 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0217 00:05:59.307087 6137 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:59.307116 6137 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 00:05:59.307204 6137 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.474649 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.478626 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.478686 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.478709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.478737 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.478754 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:01Z","lastTransitionTime":"2026-02-17T00:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.490303 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.503842 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.516193 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.538125 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.555709 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.572006 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.581656 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.581735 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.581760 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.581790 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.581811 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:01Z","lastTransitionTime":"2026-02-17T00:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.589865 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.608636 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.621841 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.684755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.684798 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.684810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.684827 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.684843 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:01Z","lastTransitionTime":"2026-02-17T00:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.787093 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.787137 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.787148 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.787166 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.787179 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:01Z","lastTransitionTime":"2026-02-17T00:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.890185 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.890237 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.890247 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.890263 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.890271 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:01Z","lastTransitionTime":"2026-02-17T00:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.895293 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns"] Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.895854 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.898701 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.898892 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.916497 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.928494 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 01:45:55.353616558 +0000 UTC Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.930706 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.944212 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.957771 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.977140 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.993424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.993477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.993490 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.993510 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.993522 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:01Z","lastTransitionTime":"2026-02-17T00:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:01 crc kubenswrapper[4690]: I0217 00:06:01.996674 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.011549 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.025466 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.039013 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.039084 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.039113 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.039138 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvjlh\" (UniqueName: \"kubernetes.io/projected/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-kube-api-access-tvjlh\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.042545 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.063647 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:57Z\\\",\\\"message\\\":\\\"0:05:57.630446 6010 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:57.631158 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:57.631182 6010 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:57.633321 6010 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 00:05:57.633334 6010 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 00:05:57.633385 6010 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0217 00:05:57.633396 6010 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 00:05:57.633413 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:57.633416 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:57.633409 6010 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 00:05:57.633435 6010 factory.go:656] Stopping watch factory\\\\nI0217 00:05:57.633445 6010 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:57.633458 6010 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:57.633465 6010 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:57.633480 6010 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0217 00:05:57.633479 6010 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:59Z\\\",\\\"message\\\":\\\"190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 00:05:59.305970 6137 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 00:05:59.305982 6137 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:59.305986 6137 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:59.305984 6137 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:59.305987 6137 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:59.306007 6137 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:59.306014 6137 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:59.306022 6137 factory.go:656] Stopping watch factory\\\\nI0217 00:05:59.306023 6137 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 00:05:59.306032 6137 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:59.306050 6137 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0217 00:05:59.306958 6137 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0217 00:05:59.307056 6137 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0217 00:05:59.307087 6137 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:59.307116 6137 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 00:05:59.307204 6137 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.079836 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.092071 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.095990 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.096035 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.096050 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.096070 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.096088 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:02Z","lastTransitionTime":"2026-02-17T00:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.105865 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.119267 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.130711 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.140279 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.140346 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.140393 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.140417 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvjlh\" (UniqueName: \"kubernetes.io/projected/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-kube-api-access-tvjlh\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.140965 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.141149 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.151893 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.157678 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvjlh\" (UniqueName: \"kubernetes.io/projected/f04a05d6-a6e4-42ac-bb58-99be30f6a38a-kube-api-access-tvjlh\") pod \"ovnkube-control-plane-749d76644c-qw6ns\" (UID: \"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.198339 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.198406 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.198418 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.198437 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.198449 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:02Z","lastTransitionTime":"2026-02-17T00:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.210933 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.301352 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.301430 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.301445 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.301465 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.301480 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:02Z","lastTransitionTime":"2026-02-17T00:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.321717 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" event={"ID":"f04a05d6-a6e4-42ac-bb58-99be30f6a38a","Type":"ContainerStarted","Data":"eec3e21f46b410e8c23d0ed9c04bf223adcebf19fba5aa0f2427d6d148b909e1"} Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.405387 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.405426 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.405436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.405453 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.405466 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:02Z","lastTransitionTime":"2026-02-17T00:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.508646 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.508951 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.508964 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.508985 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.508996 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:02Z","lastTransitionTime":"2026-02-17T00:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.611347 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.611415 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.611430 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.611450 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.611464 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:02Z","lastTransitionTime":"2026-02-17T00:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.714071 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.714117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.714133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.714154 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.714169 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:02Z","lastTransitionTime":"2026-02-17T00:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.817173 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.817230 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.817246 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.817271 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.817287 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:02Z","lastTransitionTime":"2026-02-17T00:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.919846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.919901 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.919922 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.919942 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.919956 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:02Z","lastTransitionTime":"2026-02-17T00:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.929383 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 10:06:36.154928551 +0000 UTC Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.950936 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.951071 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.951108 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.951168 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:06:18.951141187 +0000 UTC m=+52.686853948 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.951217 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.951250 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.951267 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.951266 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.951333 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.951281 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.951377 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:18.951334032 +0000 UTC m=+52.687046783 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.951618 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:18.951579139 +0000 UTC m=+52.687291960 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.951665 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:18.951651141 +0000 UTC m=+52.687364012 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.975726 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.975768 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:02 crc kubenswrapper[4690]: I0217 00:06:02.975787 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.975871 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.976012 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:02 crc kubenswrapper[4690]: E0217 00:06:02.976148 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.022741 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.022773 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.022781 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.022798 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.022810 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.023454 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-lz646"] Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.023866 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.023918 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.037330 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.052180 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.052436 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.052474 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.052495 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.052570 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:19.052547234 +0000 UTC m=+52.788260035 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.056111 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.069689 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.087185 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.107998 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.125119 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.125185 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.125209 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.125241 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.125265 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.126408 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.142251 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.152750 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.152804 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7982r\" (UniqueName: \"kubernetes.io/projected/9050c3d5-4d74-4b57-afba-1dd177ce7983-kube-api-access-7982r\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.164198 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:57Z\\\",\\\"message\\\":\\\"0:05:57.630446 6010 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:57.631158 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:57.631182 6010 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:57.633321 6010 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 00:05:57.633334 6010 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 00:05:57.633385 6010 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0217 00:05:57.633396 6010 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 00:05:57.633413 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:57.633416 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:57.633409 6010 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 00:05:57.633435 6010 factory.go:656] Stopping watch factory\\\\nI0217 00:05:57.633445 6010 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:57.633458 6010 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:57.633465 6010 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:57.633480 6010 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0217 00:05:57.633479 6010 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:59Z\\\",\\\"message\\\":\\\"190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 00:05:59.305970 6137 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 00:05:59.305982 6137 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:59.305986 6137 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:59.305984 6137 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:59.305987 6137 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:59.306007 6137 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:59.306014 6137 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:59.306022 6137 factory.go:656] Stopping watch factory\\\\nI0217 00:05:59.306023 6137 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 00:05:59.306032 6137 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:59.306050 6137 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0217 00:05:59.306958 6137 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0217 00:05:59.307056 6137 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0217 00:05:59.307087 6137 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:59.307116 6137 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 00:05:59.307204 6137 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.188649 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.204894 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.219608 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.227955 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.228042 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.228059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.228088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.228100 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.230903 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.241777 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.252563 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.254289 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.254347 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7982r\" (UniqueName: \"kubernetes.io/projected/9050c3d5-4d74-4b57-afba-1dd177ce7983-kube-api-access-7982r\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.254558 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.254638 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs podName:9050c3d5-4d74-4b57-afba-1dd177ce7983 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:03.754616598 +0000 UTC m=+37.490329349 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs") pod "network-metrics-daemon-lz646" (UID: "9050c3d5-4d74-4b57-afba-1dd177ce7983") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.263268 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.271009 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7982r\" (UniqueName: \"kubernetes.io/projected/9050c3d5-4d74-4b57-afba-1dd177ce7983-kube-api-access-7982r\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.271513 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.288939 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.288973 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.288981 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.288999 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.289008 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.299440 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.303015 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.303069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.303082 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.303103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.303116 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.315095 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.318407 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.318444 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.318457 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.318476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.318488 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.326463 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" event={"ID":"f04a05d6-a6e4-42ac-bb58-99be30f6a38a","Type":"ContainerStarted","Data":"2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.326513 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" event={"ID":"f04a05d6-a6e4-42ac-bb58-99be30f6a38a","Type":"ContainerStarted","Data":"aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da"} Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.329821 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.332653 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.332676 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.332684 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.332696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.332706 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.339515 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.345042 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.351791 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.351823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.351834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.351849 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.351859 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.362827 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:57Z\\\",\\\"message\\\":\\\"0:05:57.630446 6010 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:57.631158 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:57.631182 6010 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:57.633321 6010 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 00:05:57.633334 6010 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 00:05:57.633385 6010 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0217 00:05:57.633396 6010 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 00:05:57.633413 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:57.633416 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:57.633409 6010 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 00:05:57.633435 6010 factory.go:656] Stopping watch factory\\\\nI0217 00:05:57.633445 6010 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:57.633458 6010 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:57.633465 6010 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:57.633480 6010 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0217 00:05:57.633479 6010 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:59Z\\\",\\\"message\\\":\\\"190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 00:05:59.305970 6137 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 00:05:59.305982 6137 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:59.305986 6137 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:59.305984 6137 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:59.305987 6137 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:59.306007 6137 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:59.306014 6137 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:59.306022 6137 factory.go:656] Stopping watch factory\\\\nI0217 00:05:59.306023 6137 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 00:05:59.306032 6137 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:59.306050 6137 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0217 00:05:59.306958 6137 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0217 00:05:59.307056 6137 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0217 00:05:59.307087 6137 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:59.307116 6137 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 00:05:59.307204 6137 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.365771 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.365877 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.367528 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.367556 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.367567 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.367586 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.367605 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.376890 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.394778 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.409955 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.421106 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.433869 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.443512 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.456436 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.470731 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.470758 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.470766 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.470781 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.470790 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.473777 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.488893 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.501843 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.517525 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.534756 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.547487 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.559221 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:03Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.573170 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.573201 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.573214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.573231 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.573244 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.675338 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.675444 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.675471 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.675499 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.675518 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.761023 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.761221 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:03 crc kubenswrapper[4690]: E0217 00:06:03.761318 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs podName:9050c3d5-4d74-4b57-afba-1dd177ce7983 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:04.761295733 +0000 UTC m=+38.497008544 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs") pod "network-metrics-daemon-lz646" (UID: "9050c3d5-4d74-4b57-afba-1dd177ce7983") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.778393 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.778442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.778452 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.778465 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.778476 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.881771 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.881830 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.881848 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.881876 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.881895 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.929644 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 01:19:03.256520548 +0000 UTC Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.984782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.984852 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.984877 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.984907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:03 crc kubenswrapper[4690]: I0217 00:06:03.984934 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:03Z","lastTransitionTime":"2026-02-17T00:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.088048 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.088130 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.088142 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.088162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.088180 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:04Z","lastTransitionTime":"2026-02-17T00:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.191834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.191893 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.191905 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.191924 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.191937 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:04Z","lastTransitionTime":"2026-02-17T00:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.294736 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.294822 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.294857 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.294888 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.294912 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:04Z","lastTransitionTime":"2026-02-17T00:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.397310 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.397407 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.397423 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.397448 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.397461 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:04Z","lastTransitionTime":"2026-02-17T00:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.499956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.500015 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.500031 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.500056 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.500073 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:04Z","lastTransitionTime":"2026-02-17T00:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.603617 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.603694 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.603718 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.603750 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.603772 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:04Z","lastTransitionTime":"2026-02-17T00:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.707069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.707140 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.707157 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.707186 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.707204 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:04Z","lastTransitionTime":"2026-02-17T00:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.773139 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:04 crc kubenswrapper[4690]: E0217 00:06:04.773479 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:04 crc kubenswrapper[4690]: E0217 00:06:04.773578 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs podName:9050c3d5-4d74-4b57-afba-1dd177ce7983 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:06.773552236 +0000 UTC m=+40.509265017 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs") pod "network-metrics-daemon-lz646" (UID: "9050c3d5-4d74-4b57-afba-1dd177ce7983") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.809994 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.810036 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.810047 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.810062 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.810074 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:04Z","lastTransitionTime":"2026-02-17T00:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.912671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.912753 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.912778 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.912811 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.912835 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:04Z","lastTransitionTime":"2026-02-17T00:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.929824 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 00:03:29.480057407 +0000 UTC Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.975480 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.975532 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.975488 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:04 crc kubenswrapper[4690]: E0217 00:06:04.975683 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:04 crc kubenswrapper[4690]: I0217 00:06:04.975801 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:04 crc kubenswrapper[4690]: E0217 00:06:04.975989 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:04 crc kubenswrapper[4690]: E0217 00:06:04.976143 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:04 crc kubenswrapper[4690]: E0217 00:06:04.976386 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.019251 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.019299 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.019307 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.019323 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.019333 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:05Z","lastTransitionTime":"2026-02-17T00:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.122204 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.122256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.122268 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.122289 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.122304 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:05Z","lastTransitionTime":"2026-02-17T00:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.225289 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.225328 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.225339 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.225367 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.225378 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:05Z","lastTransitionTime":"2026-02-17T00:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.328826 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.328903 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.328927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.328960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.328983 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:05Z","lastTransitionTime":"2026-02-17T00:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.432622 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.432682 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.432751 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.432777 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.432834 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:05Z","lastTransitionTime":"2026-02-17T00:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.536601 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.536658 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.536671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.536693 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.536704 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:05Z","lastTransitionTime":"2026-02-17T00:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.639562 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.639627 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.639644 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.639669 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.639686 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:05Z","lastTransitionTime":"2026-02-17T00:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.742269 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.742416 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.742435 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.742454 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.742466 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:05Z","lastTransitionTime":"2026-02-17T00:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.845248 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.845307 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.845323 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.845348 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.845419 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:05Z","lastTransitionTime":"2026-02-17T00:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.929992 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 07:31:51.162547873 +0000 UTC Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.948077 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.948138 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.948150 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.948168 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:05 crc kubenswrapper[4690]: I0217 00:06:05.948179 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:05Z","lastTransitionTime":"2026-02-17T00:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.050134 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.050215 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.050235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.050258 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.050272 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:06Z","lastTransitionTime":"2026-02-17T00:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.153479 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.153510 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.153519 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.153532 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.153540 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:06Z","lastTransitionTime":"2026-02-17T00:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.256554 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.256582 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.256590 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.256604 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.256612 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:06Z","lastTransitionTime":"2026-02-17T00:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.359835 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.359879 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.359888 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.359907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.359917 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:06Z","lastTransitionTime":"2026-02-17T00:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.462986 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.463041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.463054 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.463071 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.463084 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:06Z","lastTransitionTime":"2026-02-17T00:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.566770 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.566827 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.566839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.566867 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.566878 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:06Z","lastTransitionTime":"2026-02-17T00:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.669873 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.669942 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.669959 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.669984 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.670001 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:06Z","lastTransitionTime":"2026-02-17T00:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.772997 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.773038 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.773069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.773088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.773100 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:06Z","lastTransitionTime":"2026-02-17T00:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.795111 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:06 crc kubenswrapper[4690]: E0217 00:06:06.795352 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:06 crc kubenswrapper[4690]: E0217 00:06:06.795524 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs podName:9050c3d5-4d74-4b57-afba-1dd177ce7983 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:10.795487099 +0000 UTC m=+44.531199910 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs") pod "network-metrics-daemon-lz646" (UID: "9050c3d5-4d74-4b57-afba-1dd177ce7983") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.876166 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.876233 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.876255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.876279 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.876296 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:06Z","lastTransitionTime":"2026-02-17T00:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.931186 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 08:57:38.403484181 +0000 UTC Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.975055 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:06 crc kubenswrapper[4690]: E0217 00:06:06.975276 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.975417 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.975485 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.975450 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:06 crc kubenswrapper[4690]: E0217 00:06:06.975724 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:06 crc kubenswrapper[4690]: E0217 00:06:06.975902 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:06 crc kubenswrapper[4690]: E0217 00:06:06.976003 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.982753 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.982846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.982871 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.982906 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.982930 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:06Z","lastTransitionTime":"2026-02-17T00:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:06 crc kubenswrapper[4690]: I0217 00:06:06.998524 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:06Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.018644 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.043653 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.060272 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.075033 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.086264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.086306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.086318 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.086335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.086348 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:07Z","lastTransitionTime":"2026-02-17T00:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.089953 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.106987 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.129388 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.150774 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.164735 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.177335 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.189397 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.189457 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.189469 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.189489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.189505 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:07Z","lastTransitionTime":"2026-02-17T00:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.196386 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b1f31929c6c1f39440ece9f4be15984a6669f4a87bf46170972067f502d95f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:57Z\\\",\\\"message\\\":\\\"0:05:57.630446 6010 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:57.631158 6010 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:57.631182 6010 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:57.633321 6010 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 00:05:57.633334 6010 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 00:05:57.633385 6010 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0217 00:05:57.633396 6010 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 00:05:57.633413 6010 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:57.633416 6010 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:57.633409 6010 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 00:05:57.633435 6010 factory.go:656] Stopping watch factory\\\\nI0217 00:05:57.633445 6010 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:57.633458 6010 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:57.633465 6010 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:57.633480 6010 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0217 00:05:57.633479 6010 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:59Z\\\",\\\"message\\\":\\\"190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 00:05:59.305970 6137 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 00:05:59.305982 6137 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:59.305986 6137 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:59.305984 6137 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:59.305987 6137 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:59.306007 6137 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:59.306014 6137 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:59.306022 6137 factory.go:656] Stopping watch factory\\\\nI0217 00:05:59.306023 6137 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 00:05:59.306032 6137 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:59.306050 6137 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0217 00:05:59.306958 6137 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0217 00:05:59.307056 6137 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0217 00:05:59.307087 6137 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:59.307116 6137 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 00:05:59.307204 6137 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.212654 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.230569 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.247521 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.262154 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:07Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.292613 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.292672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.292695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.292725 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.292750 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:07Z","lastTransitionTime":"2026-02-17T00:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.396103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.396162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.396180 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.396208 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.396232 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:07Z","lastTransitionTime":"2026-02-17T00:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.498723 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.498800 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.498828 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.498900 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.498925 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:07Z","lastTransitionTime":"2026-02-17T00:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.601642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.601709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.601729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.601755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.601774 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:07Z","lastTransitionTime":"2026-02-17T00:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.703933 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.704007 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.704025 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.704050 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.704069 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:07Z","lastTransitionTime":"2026-02-17T00:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.806582 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.806647 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.806665 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.806694 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.806713 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:07Z","lastTransitionTime":"2026-02-17T00:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.909606 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.910321 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.910342 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.910402 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.910429 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:07Z","lastTransitionTime":"2026-02-17T00:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:07 crc kubenswrapper[4690]: I0217 00:06:07.931917 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 11:27:17.534579757 +0000 UTC Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.012920 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.012989 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.013010 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.013035 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.013051 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:08Z","lastTransitionTime":"2026-02-17T00:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.116787 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.116862 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.116880 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.116907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.116925 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:08Z","lastTransitionTime":"2026-02-17T00:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.223054 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.223111 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.223124 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.223149 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.223168 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:08Z","lastTransitionTime":"2026-02-17T00:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.326036 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.326087 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.326121 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.326140 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.326152 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:08Z","lastTransitionTime":"2026-02-17T00:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.429757 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.429812 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.429825 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.429847 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.429861 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:08Z","lastTransitionTime":"2026-02-17T00:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.532938 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.532989 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.533001 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.533021 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.533032 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:08Z","lastTransitionTime":"2026-02-17T00:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.636757 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.636822 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.636841 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.636866 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.636886 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:08Z","lastTransitionTime":"2026-02-17T00:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.739782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.739852 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.739880 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.739912 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.739938 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:08Z","lastTransitionTime":"2026-02-17T00:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.842559 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.842605 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.842615 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.842631 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.842645 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:08Z","lastTransitionTime":"2026-02-17T00:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.932546 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 20:32:08.678564383 +0000 UTC Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.945387 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.945426 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.945437 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.945469 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.945480 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:08Z","lastTransitionTime":"2026-02-17T00:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.975274 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.975305 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.975452 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:08 crc kubenswrapper[4690]: I0217 00:06:08.975527 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:08 crc kubenswrapper[4690]: E0217 00:06:08.975532 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:08 crc kubenswrapper[4690]: E0217 00:06:08.975644 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:08 crc kubenswrapper[4690]: E0217 00:06:08.975846 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:08 crc kubenswrapper[4690]: E0217 00:06:08.976025 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.048489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.048549 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.048566 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.048589 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.048606 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:09Z","lastTransitionTime":"2026-02-17T00:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.150766 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.150810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.150823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.150838 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.150851 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:09Z","lastTransitionTime":"2026-02-17T00:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.253446 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.253625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.253655 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.253677 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.253692 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:09Z","lastTransitionTime":"2026-02-17T00:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.357798 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.357850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.357884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.357902 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.357911 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:09Z","lastTransitionTime":"2026-02-17T00:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.460009 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.460063 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.460079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.460108 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.460129 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:09Z","lastTransitionTime":"2026-02-17T00:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.563886 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.563924 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.563936 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.563954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.563964 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:09Z","lastTransitionTime":"2026-02-17T00:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.666610 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.666668 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.666686 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.666709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.666732 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:09Z","lastTransitionTime":"2026-02-17T00:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.772842 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.772909 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.772931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.772961 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.772982 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:09Z","lastTransitionTime":"2026-02-17T00:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.877079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.877140 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.877162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.877188 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.877205 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:09Z","lastTransitionTime":"2026-02-17T00:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.932722 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 01:43:45.864907159 +0000 UTC Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.979330 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.979425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.979438 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.979455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:09 crc kubenswrapper[4690]: I0217 00:06:09.979469 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:09Z","lastTransitionTime":"2026-02-17T00:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.082742 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.082809 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.082835 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.082866 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.082890 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:10Z","lastTransitionTime":"2026-02-17T00:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.186281 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.186349 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.186423 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.186452 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.186473 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:10Z","lastTransitionTime":"2026-02-17T00:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.289169 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.289269 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.289344 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.289424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.289450 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:10Z","lastTransitionTime":"2026-02-17T00:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.392236 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.392314 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.392338 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.392404 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.392436 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:10Z","lastTransitionTime":"2026-02-17T00:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.495747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.495796 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.495815 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.495839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.495856 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:10Z","lastTransitionTime":"2026-02-17T00:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.597707 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.597754 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.597767 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.597784 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.597796 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:10Z","lastTransitionTime":"2026-02-17T00:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.699844 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.699883 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.699894 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.699912 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.699925 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:10Z","lastTransitionTime":"2026-02-17T00:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.802675 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.802717 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.802729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.802750 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.802762 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:10Z","lastTransitionTime":"2026-02-17T00:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.838264 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:10 crc kubenswrapper[4690]: E0217 00:06:10.838525 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:10 crc kubenswrapper[4690]: E0217 00:06:10.838648 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs podName:9050c3d5-4d74-4b57-afba-1dd177ce7983 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:18.838618671 +0000 UTC m=+52.574331462 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs") pod "network-metrics-daemon-lz646" (UID: "9050c3d5-4d74-4b57-afba-1dd177ce7983") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.905525 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.905588 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.905611 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.905644 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.905664 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:10Z","lastTransitionTime":"2026-02-17T00:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.933083 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 21:37:04.95882622 +0000 UTC Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.975910 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.976026 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:10 crc kubenswrapper[4690]: E0217 00:06:10.976097 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.976199 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:10 crc kubenswrapper[4690]: E0217 00:06:10.976199 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:10 crc kubenswrapper[4690]: I0217 00:06:10.976261 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:10 crc kubenswrapper[4690]: E0217 00:06:10.976487 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:10 crc kubenswrapper[4690]: E0217 00:06:10.976616 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.008416 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.008463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.008472 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.008486 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.008494 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:11Z","lastTransitionTime":"2026-02-17T00:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.110732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.110782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.110793 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.110808 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.110887 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:11Z","lastTransitionTime":"2026-02-17T00:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.213014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.213059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.213069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.213086 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.213097 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:11Z","lastTransitionTime":"2026-02-17T00:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.315931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.315986 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.315999 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.316020 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.316032 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:11Z","lastTransitionTime":"2026-02-17T00:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.419550 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.419618 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.419642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.419672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.419694 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:11Z","lastTransitionTime":"2026-02-17T00:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.522961 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.523024 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.523042 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.523067 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.523085 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:11Z","lastTransitionTime":"2026-02-17T00:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.625333 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.625403 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.625414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.625431 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.625443 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:11Z","lastTransitionTime":"2026-02-17T00:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.727534 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.727568 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.727579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.727595 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.727606 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:11Z","lastTransitionTime":"2026-02-17T00:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.829891 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.829959 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.829973 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.829993 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.830011 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:11Z","lastTransitionTime":"2026-02-17T00:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.933721 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 12:40:34.947847826 +0000 UTC Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.934314 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.934385 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.934405 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.934432 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:11 crc kubenswrapper[4690]: I0217 00:06:11.934450 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:11Z","lastTransitionTime":"2026-02-17T00:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.009669 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.011193 4690 scope.go:117] "RemoveContainer" containerID="520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.027994 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.038082 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.038470 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.038489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.038515 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.038532 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:12Z","lastTransitionTime":"2026-02-17T00:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.047703 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.062407 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.077625 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.093485 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.110624 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.127259 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.142644 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.142689 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.142700 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.142719 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.142732 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:12Z","lastTransitionTime":"2026-02-17T00:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.143784 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.160208 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.176423 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.206082 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:59Z\\\",\\\"message\\\":\\\"190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 00:05:59.305970 6137 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 00:05:59.305982 6137 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:59.305986 6137 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:59.305984 6137 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:59.305987 6137 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:59.306007 6137 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:59.306014 6137 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:59.306022 6137 factory.go:656] Stopping watch factory\\\\nI0217 00:05:59.306023 6137 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 00:05:59.306032 6137 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:59.306050 6137 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0217 00:05:59.306958 6137 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0217 00:05:59.307056 6137 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0217 00:05:59.307087 6137 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:59.307116 6137 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 00:05:59.307204 6137 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.229584 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.245219 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.245257 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.245271 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.245290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.245304 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:12Z","lastTransitionTime":"2026-02-17T00:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.250991 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.271762 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.285703 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.299529 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.348127 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.348183 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.348202 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.348229 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.348248 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:12Z","lastTransitionTime":"2026-02-17T00:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.366067 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/1.log" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.370185 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.371515 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.388529 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.424618 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:59Z\\\",\\\"message\\\":\\\"190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 00:05:59.305970 6137 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 00:05:59.305982 6137 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:59.305986 6137 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:59.305984 6137 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:59.305987 6137 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:59.306007 6137 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:59.306014 6137 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:59.306022 6137 factory.go:656] Stopping watch factory\\\\nI0217 00:05:59.306023 6137 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 00:05:59.306032 6137 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:59.306050 6137 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0217 00:05:59.306958 6137 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0217 00:05:59.307056 6137 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0217 00:05:59.307087 6137 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:59.307116 6137 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 00:05:59.307204 6137 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.451040 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.451098 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.451114 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.451140 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.451156 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:12Z","lastTransitionTime":"2026-02-17T00:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.458233 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.479447 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.504915 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.541641 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.554477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.554552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.554582 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.554614 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.554640 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:12Z","lastTransitionTime":"2026-02-17T00:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.557505 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.570196 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.592324 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.607503 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.624961 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.638518 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.656557 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.657132 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.657194 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.657213 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.657236 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.657251 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:12Z","lastTransitionTime":"2026-02-17T00:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.670975 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.687954 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.702441 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.760074 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.760175 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.760194 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.760232 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.760263 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:12Z","lastTransitionTime":"2026-02-17T00:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.863506 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.863565 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.863580 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.863598 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.863612 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:12Z","lastTransitionTime":"2026-02-17T00:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.934098 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 22:25:51.206349973 +0000 UTC Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.966692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.966739 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.966752 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.966771 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.966784 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:12Z","lastTransitionTime":"2026-02-17T00:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.975015 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:12 crc kubenswrapper[4690]: E0217 00:06:12.975113 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.975108 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:12 crc kubenswrapper[4690]: E0217 00:06:12.975192 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.975220 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:12 crc kubenswrapper[4690]: E0217 00:06:12.975262 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:12 crc kubenswrapper[4690]: I0217 00:06:12.975400 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:12 crc kubenswrapper[4690]: E0217 00:06:12.975465 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.069634 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.069674 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.069683 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.069696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.069706 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.172944 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.172992 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.173004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.173024 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.173035 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.276462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.276512 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.276524 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.276541 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.276553 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.374798 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/2.log" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.375917 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/1.log" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.378559 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.378608 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.378630 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.378658 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.378690 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.379822 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d" exitCode=1 Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.379883 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.379944 4690 scope.go:117] "RemoveContainer" containerID="520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.380415 4690 scope.go:117] "RemoveContainer" containerID="230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d" Feb 17 00:06:13 crc kubenswrapper[4690]: E0217 00:06:13.380566 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.398061 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.398123 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.398146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.398184 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.398207 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.407900 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: E0217 00:06:13.425904 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.428249 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.431106 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.431162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.431181 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.431206 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.431224 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.445807 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: E0217 00:06:13.448351 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.452618 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.452660 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.452671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.452688 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.452700 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.461908 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: E0217 00:06:13.485246 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.493751 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.493791 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.493803 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.493821 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.493833 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.507744 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://520b1008983ee76a5bdf311567eb156605b9e101d6026af087381d1defc9b87f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:05:59Z\\\",\\\"message\\\":\\\"190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 00:05:59.305970 6137 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 00:05:59.305982 6137 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 00:05:59.305986 6137 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 00:05:59.305984 6137 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 00:05:59.305987 6137 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0217 00:05:59.306007 6137 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 00:05:59.306014 6137 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 00:05:59.306022 6137 factory.go:656] Stopping watch factory\\\\nI0217 00:05:59.306023 6137 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 00:05:59.306032 6137 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 00:05:59.306050 6137 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0217 00:05:59.306958 6137 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0217 00:05:59.307056 6137 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0217 00:05:59.307087 6137 ovnkube.go:599] Stopped ovnkube\\\\nI0217 00:05:59.307116 6137 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 00:05:59.307204 6137 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"1 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:7FOCZ3GVMQ1pwQKJahWmE09uJDRx6ab8xxcEYE] map[] [{operators.coreos.com/v1alpha1 CatalogSource certified-operators faf75094-01cc-4ebf-8761-1361fa2bf31e 0xc0076f7a2d 0xc0076f7a2e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: certified-operators,olm.managed: true,},ClusterIP:10.217.5.214,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.214],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0217 00:06:12.996747 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: E0217 00:06:13.514983 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.520695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.520733 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.520744 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.520760 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.520772 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.531188 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: E0217 00:06:13.535424 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: E0217 00:06:13.535722 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.537447 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.537494 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.537511 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.537532 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.537547 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.548520 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.561150 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.571951 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.581116 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.589627 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.599059 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.614504 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.625764 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.636758 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.639500 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.639541 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.639552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.639572 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.639585 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.647923 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.742836 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.742889 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.742907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.742930 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.742947 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.845237 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.845303 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.845320 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.845345 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.845392 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.934635 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 02:01:37.020097332 +0000 UTC Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.948230 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.948290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.948309 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.948336 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:13 crc kubenswrapper[4690]: I0217 00:06:13.948354 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:13Z","lastTransitionTime":"2026-02-17T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.051125 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.051184 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.051203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.051226 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.051241 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:14Z","lastTransitionTime":"2026-02-17T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.154803 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.154877 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.154894 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.154920 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.154938 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:14Z","lastTransitionTime":"2026-02-17T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.257682 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.257781 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.257808 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.257841 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.257863 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:14Z","lastTransitionTime":"2026-02-17T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.360416 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.360506 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.360519 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.360557 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.360571 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:14Z","lastTransitionTime":"2026-02-17T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.386154 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/2.log" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.390810 4690 scope.go:117] "RemoveContainer" containerID="230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d" Feb 17 00:06:14 crc kubenswrapper[4690]: E0217 00:06:14.390990 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.412469 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.425985 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.440953 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.462459 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"1 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:7FOCZ3GVMQ1pwQKJahWmE09uJDRx6ab8xxcEYE] map[] [{operators.coreos.com/v1alpha1 CatalogSource certified-operators faf75094-01cc-4ebf-8761-1361fa2bf31e 0xc0076f7a2d 0xc0076f7a2e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: certified-operators,olm.managed: true,},ClusterIP:10.217.5.214,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.214],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0217 00:06:12.996747 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.463759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.463806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.463820 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.463840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.463853 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:14Z","lastTransitionTime":"2026-02-17T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.480220 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.501900 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.518696 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.534647 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.547438 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.561283 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.566926 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.566977 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.566994 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.567018 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.567038 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:14Z","lastTransitionTime":"2026-02-17T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.577624 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.593453 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.611890 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.631268 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.646935 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.666113 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.670418 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.670631 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.670771 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.670912 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.671044 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:14Z","lastTransitionTime":"2026-02-17T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.774749 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.774810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.774828 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.774855 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.774873 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:14Z","lastTransitionTime":"2026-02-17T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.877907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.877934 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.877941 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.877954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.877962 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:14Z","lastTransitionTime":"2026-02-17T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.935686 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 18:11:51.62411159 +0000 UTC Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.975079 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.975545 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.975749 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.975855 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:14 crc kubenswrapper[4690]: E0217 00:06:14.975748 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:14 crc kubenswrapper[4690]: E0217 00:06:14.976215 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:14 crc kubenswrapper[4690]: E0217 00:06:14.976462 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:14 crc kubenswrapper[4690]: E0217 00:06:14.976587 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.980344 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.980428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.980446 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.980471 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:14 crc kubenswrapper[4690]: I0217 00:06:14.980489 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:14Z","lastTransitionTime":"2026-02-17T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.082854 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.082889 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.082897 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.082917 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.082926 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:15Z","lastTransitionTime":"2026-02-17T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.185147 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.185215 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.185234 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.185260 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.185277 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:15Z","lastTransitionTime":"2026-02-17T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.287784 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.287823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.287834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.287852 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.287865 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:15Z","lastTransitionTime":"2026-02-17T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.390471 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.390502 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.390513 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.390528 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.390566 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:15Z","lastTransitionTime":"2026-02-17T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.493576 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.493642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.493664 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.493695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.493720 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:15Z","lastTransitionTime":"2026-02-17T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.595991 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.596043 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.596054 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.596074 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.596088 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:15Z","lastTransitionTime":"2026-02-17T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.698550 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.698599 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.698618 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.698643 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.698664 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:15Z","lastTransitionTime":"2026-02-17T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.800690 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.800757 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.800767 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.800781 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.800791 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:15Z","lastTransitionTime":"2026-02-17T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.903738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.903797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.903811 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.903830 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.903844 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:15Z","lastTransitionTime":"2026-02-17T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:15 crc kubenswrapper[4690]: I0217 00:06:15.936445 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 22:12:22.190658256 +0000 UTC Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.006281 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.006332 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.006343 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.006386 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.006400 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:16Z","lastTransitionTime":"2026-02-17T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.108607 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.108645 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.108653 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.108667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.108675 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:16Z","lastTransitionTime":"2026-02-17T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.212539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.212615 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.212638 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.212711 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.212734 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:16Z","lastTransitionTime":"2026-02-17T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.316177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.316239 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.316264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.316292 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.316313 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:16Z","lastTransitionTime":"2026-02-17T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.418419 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.419025 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.419171 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.419386 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.419582 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:16Z","lastTransitionTime":"2026-02-17T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.523280 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.523332 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.523344 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.523376 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.523392 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:16Z","lastTransitionTime":"2026-02-17T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.626537 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.626588 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.626597 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.626615 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.626624 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:16Z","lastTransitionTime":"2026-02-17T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.729188 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.729240 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.729258 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.729281 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.729300 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:16Z","lastTransitionTime":"2026-02-17T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.832305 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.832409 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.832429 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.832456 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.832473 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:16Z","lastTransitionTime":"2026-02-17T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.935673 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.935721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.935732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.935750 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.935764 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:16Z","lastTransitionTime":"2026-02-17T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.936980 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 08:23:01.92880692 +0000 UTC Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.975442 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.975526 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.975585 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:16 crc kubenswrapper[4690]: E0217 00:06:16.975765 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.975839 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:16 crc kubenswrapper[4690]: E0217 00:06:16.975905 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:16 crc kubenswrapper[4690]: E0217 00:06:16.976017 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:16 crc kubenswrapper[4690]: E0217 00:06:16.976201 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:16 crc kubenswrapper[4690]: I0217 00:06:16.996313 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.011400 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.028781 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.038294 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.038327 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.038335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.038349 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.038372 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:17Z","lastTransitionTime":"2026-02-17T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.046683 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.064090 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.082609 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.105768 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"1 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:7FOCZ3GVMQ1pwQKJahWmE09uJDRx6ab8xxcEYE] map[] [{operators.coreos.com/v1alpha1 CatalogSource certified-operators faf75094-01cc-4ebf-8761-1361fa2bf31e 0xc0076f7a2d 0xc0076f7a2e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: certified-operators,olm.managed: true,},ClusterIP:10.217.5.214,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.214],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0217 00:06:12.996747 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.128924 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.141853 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.141892 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.141904 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.141921 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.141932 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:17Z","lastTransitionTime":"2026-02-17T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.148626 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.170193 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.187851 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.201483 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.215369 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.237335 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.244070 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.244141 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.244159 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.244184 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.244201 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:17Z","lastTransitionTime":"2026-02-17T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.255287 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.275340 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:17Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.347554 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.347611 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.347633 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.347661 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.347685 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:17Z","lastTransitionTime":"2026-02-17T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.450595 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.450652 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.450667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.450692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.450704 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:17Z","lastTransitionTime":"2026-02-17T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.553012 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.553073 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.553085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.553103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.553117 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:17Z","lastTransitionTime":"2026-02-17T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.656143 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.656211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.656229 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.656260 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.656278 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:17Z","lastTransitionTime":"2026-02-17T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.758587 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.758655 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.758679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.758709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.758733 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:17Z","lastTransitionTime":"2026-02-17T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.861387 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.861450 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.861463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.861480 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.861492 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:17Z","lastTransitionTime":"2026-02-17T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.937392 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 15:23:49.397553896 +0000 UTC Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.963558 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.963626 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.963651 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.963681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:17 crc kubenswrapper[4690]: I0217 00:06:17.963706 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:17Z","lastTransitionTime":"2026-02-17T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.066395 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.066422 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.066430 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.066444 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.066453 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:18Z","lastTransitionTime":"2026-02-17T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.169705 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.169762 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.169781 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.169806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.169823 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:18Z","lastTransitionTime":"2026-02-17T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.272466 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.272532 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.272552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.272581 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.272605 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:18Z","lastTransitionTime":"2026-02-17T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.376161 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.376228 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.376245 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.376269 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.376290 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:18Z","lastTransitionTime":"2026-02-17T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.479281 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.479312 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.479321 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.479335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.479345 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:18Z","lastTransitionTime":"2026-02-17T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.582124 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.582159 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.582168 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.582201 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.582213 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:18Z","lastTransitionTime":"2026-02-17T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.685494 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.685575 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.685600 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.685662 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.685687 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:18Z","lastTransitionTime":"2026-02-17T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.788818 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.788862 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.788878 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.788901 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.788919 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:18Z","lastTransitionTime":"2026-02-17T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.892275 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.892328 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.892351 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.892424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.892447 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:18Z","lastTransitionTime":"2026-02-17T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.935339 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:18 crc kubenswrapper[4690]: E0217 00:06:18.935648 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:18 crc kubenswrapper[4690]: E0217 00:06:18.935735 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs podName:9050c3d5-4d74-4b57-afba-1dd177ce7983 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:34.935708195 +0000 UTC m=+68.671420996 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs") pod "network-metrics-daemon-lz646" (UID: "9050c3d5-4d74-4b57-afba-1dd177ce7983") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.938532 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 22:06:52.359315066 +0000 UTC Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.975861 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.975947 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.975989 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.976033 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:18 crc kubenswrapper[4690]: E0217 00:06:18.976333 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:18 crc kubenswrapper[4690]: E0217 00:06:18.976535 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:18 crc kubenswrapper[4690]: E0217 00:06:18.976760 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:18 crc kubenswrapper[4690]: E0217 00:06:18.976861 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.994976 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.995048 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.995073 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.995102 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:18 crc kubenswrapper[4690]: I0217 00:06:18.995121 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:18Z","lastTransitionTime":"2026-02-17T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.036422 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.036529 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.036568 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.036590 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.036712 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.036726 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.036736 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.036738 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:06:51.03670149 +0000 UTC m=+84.772414281 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.036781 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:51.036772502 +0000 UTC m=+84.772485243 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.036783 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.036801 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.036881 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:51.036858375 +0000 UTC m=+84.772571136 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.036921 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:51.036892035 +0000 UTC m=+84.772604796 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.098152 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.098194 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.098203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.098248 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.098257 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:19Z","lastTransitionTime":"2026-02-17T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.137852 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.138000 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.138035 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.138047 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:19 crc kubenswrapper[4690]: E0217 00:06:19.138094 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 00:06:51.138081747 +0000 UTC m=+84.873794498 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.201406 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.201480 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.201495 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.201510 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.201520 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:19Z","lastTransitionTime":"2026-02-17T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.305304 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.305352 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.305399 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.305422 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.305440 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:19Z","lastTransitionTime":"2026-02-17T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.407299 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.407409 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.407436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.407463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.407485 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:19Z","lastTransitionTime":"2026-02-17T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.510440 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.510772 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.511041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.511268 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.511528 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:19Z","lastTransitionTime":"2026-02-17T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.614676 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.614719 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.614732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.614770 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.614782 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:19Z","lastTransitionTime":"2026-02-17T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.717298 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.717340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.717350 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.717395 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.717408 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:19Z","lastTransitionTime":"2026-02-17T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.826783 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.826855 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.826868 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.826886 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.826899 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:19Z","lastTransitionTime":"2026-02-17T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.929886 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.929940 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.929951 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.929969 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.929981 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:19Z","lastTransitionTime":"2026-02-17T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:19 crc kubenswrapper[4690]: I0217 00:06:19.939606 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 09:34:08.167497425 +0000 UTC Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.032979 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.033032 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.033046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.033066 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.033081 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:20Z","lastTransitionTime":"2026-02-17T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.136492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.136538 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.136548 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.136564 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.136575 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:20Z","lastTransitionTime":"2026-02-17T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.240793 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.240875 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.240897 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.240933 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.240957 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:20Z","lastTransitionTime":"2026-02-17T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.344671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.344725 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.344741 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.344766 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.344782 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:20Z","lastTransitionTime":"2026-02-17T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.446903 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.446954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.446975 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.446994 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.447008 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:20Z","lastTransitionTime":"2026-02-17T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.549801 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.549872 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.549908 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.549938 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.549960 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:20Z","lastTransitionTime":"2026-02-17T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.653071 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.653122 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.653136 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.653153 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.653165 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:20Z","lastTransitionTime":"2026-02-17T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.756071 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.756105 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.756115 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.756129 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.756137 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:20Z","lastTransitionTime":"2026-02-17T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.858533 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.858595 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.858611 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.858634 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.858652 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:20Z","lastTransitionTime":"2026-02-17T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.940480 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 10:21:01.54405922 +0000 UTC Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.961451 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.961496 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.961509 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.961527 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.961541 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:20Z","lastTransitionTime":"2026-02-17T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.975412 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.975450 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:20 crc kubenswrapper[4690]: E0217 00:06:20.975564 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.975597 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:20 crc kubenswrapper[4690]: I0217 00:06:20.975610 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:20 crc kubenswrapper[4690]: E0217 00:06:20.975703 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:20 crc kubenswrapper[4690]: E0217 00:06:20.975789 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:20 crc kubenswrapper[4690]: E0217 00:06:20.975936 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.063945 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.064251 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.064268 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.064293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.064311 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:21Z","lastTransitionTime":"2026-02-17T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.167223 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.167296 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.167318 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.167345 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.167418 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:21Z","lastTransitionTime":"2026-02-17T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.270317 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.270373 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.270385 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.270403 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.270414 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:21Z","lastTransitionTime":"2026-02-17T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.372811 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.372876 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.372898 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.372927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.372952 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:21Z","lastTransitionTime":"2026-02-17T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.475248 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.475314 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.475332 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.475394 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.475413 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:21Z","lastTransitionTime":"2026-02-17T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.578059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.578137 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.578159 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.578190 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.578219 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:21Z","lastTransitionTime":"2026-02-17T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.681146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.681209 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.681222 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.681243 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.681257 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:21Z","lastTransitionTime":"2026-02-17T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.783916 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.783981 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.783999 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.784023 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.784040 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:21Z","lastTransitionTime":"2026-02-17T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.887223 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.887286 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.887306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.887330 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.887347 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:21Z","lastTransitionTime":"2026-02-17T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.941416 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 03:12:30.312034613 +0000 UTC Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.989710 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.989764 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.989783 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.989804 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:21 crc kubenswrapper[4690]: I0217 00:06:21.989821 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:21Z","lastTransitionTime":"2026-02-17T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.094059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.094190 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.094207 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.094232 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.094253 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:22Z","lastTransitionTime":"2026-02-17T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.198062 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.198125 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.198144 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.198172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.198193 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:22Z","lastTransitionTime":"2026-02-17T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.301044 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.301109 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.301128 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.301154 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.301173 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:22Z","lastTransitionTime":"2026-02-17T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.404628 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.404688 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.404705 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.404733 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.404749 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:22Z","lastTransitionTime":"2026-02-17T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.522837 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.522904 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.522927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.522957 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.522979 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:22Z","lastTransitionTime":"2026-02-17T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.554848 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.572571 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.580285 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.597613 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.615966 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.627145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.627197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.627214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.627239 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.627257 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:22Z","lastTransitionTime":"2026-02-17T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.631591 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.646646 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.662480 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.681225 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.710624 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"1 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:7FOCZ3GVMQ1pwQKJahWmE09uJDRx6ab8xxcEYE] map[] [{operators.coreos.com/v1alpha1 CatalogSource certified-operators faf75094-01cc-4ebf-8761-1361fa2bf31e 0xc0076f7a2d 0xc0076f7a2e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: certified-operators,olm.managed: true,},ClusterIP:10.217.5.214,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.214],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0217 00:06:12.996747 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.730187 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.730494 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.730732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.730971 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.731208 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:22Z","lastTransitionTime":"2026-02-17T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.734757 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.757204 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.777407 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.792821 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.809244 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.821829 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.833822 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.833861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.833873 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.833893 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.833905 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:22Z","lastTransitionTime":"2026-02-17T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.833933 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.847787 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.936301 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.936409 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.936429 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.936453 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.936469 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:22Z","lastTransitionTime":"2026-02-17T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.942566 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 11:11:19.354722967 +0000 UTC Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.975221 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.975417 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:22 crc kubenswrapper[4690]: E0217 00:06:22.975558 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.975711 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:22 crc kubenswrapper[4690]: I0217 00:06:22.975716 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:22 crc kubenswrapper[4690]: E0217 00:06:22.975874 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:22 crc kubenswrapper[4690]: E0217 00:06:22.976251 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:22 crc kubenswrapper[4690]: E0217 00:06:22.976868 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.039028 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.039072 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.039086 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.039104 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.039117 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.141989 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.142042 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.142058 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.142080 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.142096 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.244216 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.244266 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.244282 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.244304 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.244320 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.347469 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.347524 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.347544 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.347573 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.347594 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.449396 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.449448 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.449457 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.449469 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.449479 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.552444 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.552518 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.552540 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.552566 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.552587 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.648052 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.648124 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.648146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.648175 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.648196 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: E0217 00:06:23.670270 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.675259 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.675419 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.675449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.675480 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.675503 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: E0217 00:06:23.698208 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.703275 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.703350 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.703476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.703512 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.703537 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: E0217 00:06:23.723600 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.727927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.727992 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.728021 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.728050 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.728072 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: E0217 00:06:23.753119 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.758586 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.758650 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.758669 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.758695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.758714 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: E0217 00:06:23.778811 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:23 crc kubenswrapper[4690]: E0217 00:06:23.779464 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.781542 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.781614 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.781638 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.781667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.781690 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.884893 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.884956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.884974 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.884998 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.885076 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.942757 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 17:13:34.357176856 +0000 UTC Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.988119 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.988192 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.988218 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.988248 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:23 crc kubenswrapper[4690]: I0217 00:06:23.988271 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:23Z","lastTransitionTime":"2026-02-17T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.091795 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.091834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.091846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.091862 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.091874 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:24Z","lastTransitionTime":"2026-02-17T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.197505 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.197563 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.197579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.197605 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.197624 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:24Z","lastTransitionTime":"2026-02-17T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.300772 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.300850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.300876 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.300922 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.300947 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:24Z","lastTransitionTime":"2026-02-17T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.403661 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.403747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.403770 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.403799 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.403818 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:24Z","lastTransitionTime":"2026-02-17T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.506280 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.506409 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.506447 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.506476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.506497 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:24Z","lastTransitionTime":"2026-02-17T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.609491 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.609564 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.609588 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.609676 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.609704 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:24Z","lastTransitionTime":"2026-02-17T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.712211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.712292 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.712313 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.712340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.712388 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:24Z","lastTransitionTime":"2026-02-17T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.815121 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.815183 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.815200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.815223 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.815240 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:24Z","lastTransitionTime":"2026-02-17T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.918235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.918290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.918308 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.918338 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.918405 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:24Z","lastTransitionTime":"2026-02-17T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.943003 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 23:19:17.956646748 +0000 UTC Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.975498 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.975488 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.975550 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:24 crc kubenswrapper[4690]: I0217 00:06:24.975511 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:24 crc kubenswrapper[4690]: E0217 00:06:24.975701 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:24 crc kubenswrapper[4690]: E0217 00:06:24.975859 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:24 crc kubenswrapper[4690]: E0217 00:06:24.976012 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:24 crc kubenswrapper[4690]: E0217 00:06:24.976098 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.022721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.022847 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.022914 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.022954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.023019 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:25Z","lastTransitionTime":"2026-02-17T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.125803 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.125849 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.125866 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.125892 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.125908 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:25Z","lastTransitionTime":"2026-02-17T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.229129 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.229177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.229195 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.229222 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.229240 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:25Z","lastTransitionTime":"2026-02-17T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.332143 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.332205 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.332223 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.332250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.332267 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:25Z","lastTransitionTime":"2026-02-17T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.435128 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.435201 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.435224 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.435251 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.435271 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:25Z","lastTransitionTime":"2026-02-17T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.537815 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.537910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.537946 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.537978 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.538015 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:25Z","lastTransitionTime":"2026-02-17T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.641127 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.641191 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.641201 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.641214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.641222 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:25Z","lastTransitionTime":"2026-02-17T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.744197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.744249 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.744265 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.744289 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.744306 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:25Z","lastTransitionTime":"2026-02-17T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.847971 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.848040 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.848062 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.848093 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.848116 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:25Z","lastTransitionTime":"2026-02-17T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.943787 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 12:31:40.697494165 +0000 UTC Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.950799 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.950831 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.950838 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.950850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:25 crc kubenswrapper[4690]: I0217 00:06:25.950858 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:25Z","lastTransitionTime":"2026-02-17T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.053048 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.053091 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.053102 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.053121 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.053132 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:26Z","lastTransitionTime":"2026-02-17T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.156505 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.156581 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.156606 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.156641 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.156665 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:26Z","lastTransitionTime":"2026-02-17T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.260841 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.260887 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.260904 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.260926 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.260944 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:26Z","lastTransitionTime":"2026-02-17T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.363138 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.363201 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.363222 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.363250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.363274 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:26Z","lastTransitionTime":"2026-02-17T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.465120 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.465167 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.465495 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.465557 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.465568 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:26Z","lastTransitionTime":"2026-02-17T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.572659 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.572729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.572763 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.572795 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.572815 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:26Z","lastTransitionTime":"2026-02-17T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.676569 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.676637 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.676654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.676679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.676697 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:26Z","lastTransitionTime":"2026-02-17T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.779347 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.779440 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.779458 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.779944 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.780016 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:26Z","lastTransitionTime":"2026-02-17T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.883486 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.883573 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.883591 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.883614 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.883632 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:26Z","lastTransitionTime":"2026-02-17T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.943958 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 15:35:47.080298186 +0000 UTC Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.975092 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.975135 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.975149 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.975198 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:26 crc kubenswrapper[4690]: E0217 00:06:26.975409 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:26 crc kubenswrapper[4690]: E0217 00:06:26.975511 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:26 crc kubenswrapper[4690]: E0217 00:06:26.975682 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:26 crc kubenswrapper[4690]: E0217 00:06:26.976232 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.976514 4690 scope.go:117] "RemoveContainer" containerID="230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d" Feb 17 00:06:26 crc kubenswrapper[4690]: E0217 00:06:26.976733 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.986339 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.986427 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.986456 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.986481 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.986502 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:26Z","lastTransitionTime":"2026-02-17T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:26 crc kubenswrapper[4690]: I0217 00:06:26.998484 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:26Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.018654 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.038752 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.056525 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.088656 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.088706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.088723 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.088746 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.088764 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:27Z","lastTransitionTime":"2026-02-17T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.093726 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"1 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:7FOCZ3GVMQ1pwQKJahWmE09uJDRx6ab8xxcEYE] map[] [{operators.coreos.com/v1alpha1 CatalogSource certified-operators faf75094-01cc-4ebf-8761-1361fa2bf31e 0xc0076f7a2d 0xc0076f7a2e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: certified-operators,olm.managed: true,},ClusterIP:10.217.5.214,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.214],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0217 00:06:12.996747 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.116435 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.136530 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.152772 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8057003-037a-43f0-a29f-8b30d2f9f973\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://971fe3f679431845df0be1a451c36f170cd038649d687a730a6b74adfe370b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dae842d1ef134d705b87402029edb7e1acc4d3b1efbc458eb8ceaf07f2426db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d468f64b7ca19373b20612c77a5482144dffdcd3e76d0377af6a421a4770d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.174168 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.190150 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.191168 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.191205 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.191217 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.191233 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.191245 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:27Z","lastTransitionTime":"2026-02-17T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.203423 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.218598 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.235785 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.250255 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.267269 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.286249 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.295846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.295884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.295895 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.295910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.295923 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:27Z","lastTransitionTime":"2026-02-17T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.309539 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.398949 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.399005 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.399022 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.399041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.399057 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:27Z","lastTransitionTime":"2026-02-17T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.501473 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.501526 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.501547 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.501574 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.501595 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:27Z","lastTransitionTime":"2026-02-17T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.605595 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.605659 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.605679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.605707 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.605728 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:27Z","lastTransitionTime":"2026-02-17T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.707998 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.708103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.708122 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.708146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.708163 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:27Z","lastTransitionTime":"2026-02-17T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.810030 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.810091 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.810105 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.810122 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.810133 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:27Z","lastTransitionTime":"2026-02-17T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.912434 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.912699 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.912782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.912877 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.912956 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:27Z","lastTransitionTime":"2026-02-17T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:27 crc kubenswrapper[4690]: I0217 00:06:27.945064 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 02:30:13.69955285 +0000 UTC Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.016759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.016818 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.016836 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.016861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.016878 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:28Z","lastTransitionTime":"2026-02-17T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.120744 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.120803 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.120819 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.120844 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.120863 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:28Z","lastTransitionTime":"2026-02-17T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.224339 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.224451 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.224469 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.224495 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.224512 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:28Z","lastTransitionTime":"2026-02-17T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.327708 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.327745 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.327754 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.327768 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.327777 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:28Z","lastTransitionTime":"2026-02-17T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.431012 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.431103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.431129 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.431178 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.431202 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:28Z","lastTransitionTime":"2026-02-17T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.533886 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.533935 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.533952 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.533974 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.533990 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:28Z","lastTransitionTime":"2026-02-17T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.639679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.639884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.639910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.639942 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.639967 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:28Z","lastTransitionTime":"2026-02-17T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.743313 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.743436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.743454 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.743478 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.743494 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:28Z","lastTransitionTime":"2026-02-17T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.846493 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.846549 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.846562 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.846583 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.846599 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:28Z","lastTransitionTime":"2026-02-17T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.945629 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 05:04:35.231835974 +0000 UTC Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.949617 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.949662 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.949679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.949701 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.949717 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:28Z","lastTransitionTime":"2026-02-17T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.975849 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.975849 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:28 crc kubenswrapper[4690]: E0217 00:06:28.976067 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:28 crc kubenswrapper[4690]: E0217 00:06:28.976195 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.976193 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:28 crc kubenswrapper[4690]: I0217 00:06:28.975889 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:28 crc kubenswrapper[4690]: E0217 00:06:28.976498 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:28 crc kubenswrapper[4690]: E0217 00:06:28.976768 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.052842 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.052905 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.052922 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.052946 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.052963 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:29Z","lastTransitionTime":"2026-02-17T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.155402 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.155481 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.155503 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.155541 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.155570 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:29Z","lastTransitionTime":"2026-02-17T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.258672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.258724 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.258763 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.258788 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.258805 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:29Z","lastTransitionTime":"2026-02-17T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.362003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.362040 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.362052 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.362069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.362082 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:29Z","lastTransitionTime":"2026-02-17T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.464779 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.464845 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.464866 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.464889 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.464909 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:29Z","lastTransitionTime":"2026-02-17T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.567579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.567642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.567661 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.567683 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.567700 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:29Z","lastTransitionTime":"2026-02-17T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.670842 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.670889 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.670899 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.670913 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.670925 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:29Z","lastTransitionTime":"2026-02-17T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.774127 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.774189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.774205 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.774229 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.774247 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:29Z","lastTransitionTime":"2026-02-17T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.876903 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.877002 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.877057 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.877083 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.877104 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:29Z","lastTransitionTime":"2026-02-17T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.946577 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 21:38:04.129249382 +0000 UTC Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.980039 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.980079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.980090 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.980104 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:29 crc kubenswrapper[4690]: I0217 00:06:29.980116 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:29Z","lastTransitionTime":"2026-02-17T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.082430 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.082545 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.082559 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.082572 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.082579 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:30Z","lastTransitionTime":"2026-02-17T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.186155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.186208 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.186221 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.186241 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.186253 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:30Z","lastTransitionTime":"2026-02-17T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.289652 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.289699 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.289712 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.289729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.289742 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:30Z","lastTransitionTime":"2026-02-17T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.392513 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.392548 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.392556 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.392568 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.392577 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:30Z","lastTransitionTime":"2026-02-17T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.494879 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.494935 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.494951 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.494973 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.494990 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:30Z","lastTransitionTime":"2026-02-17T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.597875 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.597938 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.597960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.597989 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.598011 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:30Z","lastTransitionTime":"2026-02-17T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.701112 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.701178 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.701199 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.701229 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.701247 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:30Z","lastTransitionTime":"2026-02-17T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.804320 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.804422 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.804445 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.804471 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.804494 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:30Z","lastTransitionTime":"2026-02-17T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.907425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.907470 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.907489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.907511 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.907526 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:30Z","lastTransitionTime":"2026-02-17T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.946831 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 05:31:27.51982574 +0000 UTC Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.975572 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:30 crc kubenswrapper[4690]: E0217 00:06:30.975722 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.975744 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.975769 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:30 crc kubenswrapper[4690]: I0217 00:06:30.975581 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:30 crc kubenswrapper[4690]: E0217 00:06:30.975821 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:30 crc kubenswrapper[4690]: E0217 00:06:30.975950 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:30 crc kubenswrapper[4690]: E0217 00:06:30.976217 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.009268 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.009316 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.009325 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.009336 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.009345 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:31Z","lastTransitionTime":"2026-02-17T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.112284 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.112325 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.112336 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.112372 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.112385 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:31Z","lastTransitionTime":"2026-02-17T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.214819 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.214859 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.214868 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.214883 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.214893 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:31Z","lastTransitionTime":"2026-02-17T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.317686 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.317756 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.317775 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.317802 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.317821 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:31Z","lastTransitionTime":"2026-02-17T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.419955 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.419993 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.420005 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.420022 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.420035 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:31Z","lastTransitionTime":"2026-02-17T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.522061 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.522102 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.522113 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.522128 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.522140 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:31Z","lastTransitionTime":"2026-02-17T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.625390 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.625441 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.625453 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.625473 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.625486 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:31Z","lastTransitionTime":"2026-02-17T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.727618 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.727654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.727664 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.727678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.727687 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:31Z","lastTransitionTime":"2026-02-17T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.830057 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.830113 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.830131 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.830155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.830172 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:31Z","lastTransitionTime":"2026-02-17T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.933765 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.933840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.933861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.933886 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.933902 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:31Z","lastTransitionTime":"2026-02-17T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:31 crc kubenswrapper[4690]: I0217 00:06:31.947221 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 06:31:13.709355663 +0000 UTC Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.036995 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.037033 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.037045 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.037061 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.037075 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:32Z","lastTransitionTime":"2026-02-17T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.139706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.139772 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.139786 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.139811 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.139824 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:32Z","lastTransitionTime":"2026-02-17T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.243708 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.243752 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.243763 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.243780 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.243791 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:32Z","lastTransitionTime":"2026-02-17T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.347576 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.347635 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.347648 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.347674 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.347695 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:32Z","lastTransitionTime":"2026-02-17T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.451070 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.451158 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.451213 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.451240 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.451281 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:32Z","lastTransitionTime":"2026-02-17T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.554212 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.554258 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.554267 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.554285 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.554294 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:32Z","lastTransitionTime":"2026-02-17T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.657243 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.657309 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.657327 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.657384 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.657403 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:32Z","lastTransitionTime":"2026-02-17T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.761136 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.761177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.761186 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.761205 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.761215 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:32Z","lastTransitionTime":"2026-02-17T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.864999 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.865048 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.865060 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.865080 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.865092 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:32Z","lastTransitionTime":"2026-02-17T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.948288 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 05:37:22.968695209 +0000 UTC Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.967460 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.967491 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.967501 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.967514 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.967524 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:32Z","lastTransitionTime":"2026-02-17T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.976107 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.976199 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.976307 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:32 crc kubenswrapper[4690]: E0217 00:06:32.976347 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:32 crc kubenswrapper[4690]: I0217 00:06:32.976465 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:32 crc kubenswrapper[4690]: E0217 00:06:32.976651 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:32 crc kubenswrapper[4690]: E0217 00:06:32.976786 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:32 crc kubenswrapper[4690]: E0217 00:06:32.976986 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.069542 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.069605 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.069622 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.069643 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.069660 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.172292 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.172348 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.172394 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.172425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.172447 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.275177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.275229 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.275247 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.275270 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.275287 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.378270 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.378307 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.378319 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.378333 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.378343 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.481604 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.481699 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.481713 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.481731 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.481745 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.585422 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.585492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.585548 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.585586 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.585606 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.690028 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.690123 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.690147 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.690179 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.690217 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.794560 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.794632 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.794653 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.794681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.794703 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.897795 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.897895 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.897923 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.897964 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.897990 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.917187 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.917261 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.917282 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.917307 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.917326 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: E0217 00:06:33.938640 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:33Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.943156 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.943397 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.943428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.943456 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.943475 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.948649 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 14:59:47.078048771 +0000 UTC Feb 17 00:06:33 crc kubenswrapper[4690]: E0217 00:06:33.964231 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:33Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.969147 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.969191 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.969210 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.969233 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.969246 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:33 crc kubenswrapper[4690]: E0217 00:06:33.982864 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:33Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.987145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.987264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.987288 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.987415 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:33 crc kubenswrapper[4690]: I0217 00:06:33.987517 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:33Z","lastTransitionTime":"2026-02-17T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: E0217 00:06:34.004002 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:34Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.007758 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.007808 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.007819 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.007836 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.007849 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: E0217 00:06:34.019539 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:34Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:34 crc kubenswrapper[4690]: E0217 00:06:34.019649 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.023041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.023063 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.023072 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.023087 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.023100 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.126047 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.126138 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.126159 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.126190 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.126214 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.229617 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.229666 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.229677 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.229692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.229703 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.332066 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.332112 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.332123 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.332145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.332157 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.434818 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.434868 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.434883 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.434902 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.434914 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.537186 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.537253 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.537264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.537291 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.537302 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.639250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.639288 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.639297 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.639315 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.639324 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.742245 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.742320 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.742337 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.742373 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.742388 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.845555 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.845633 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.845647 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.845667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.845679 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.948610 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.948651 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.948660 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.948677 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.948687 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:34Z","lastTransitionTime":"2026-02-17T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.948851 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 06:29:30.248174628 +0000 UTC Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.976054 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.976117 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:34 crc kubenswrapper[4690]: E0217 00:06:34.976162 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.976059 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:34 crc kubenswrapper[4690]: E0217 00:06:34.976286 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:34 crc kubenswrapper[4690]: E0217 00:06:34.976407 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:34 crc kubenswrapper[4690]: I0217 00:06:34.977089 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:34 crc kubenswrapper[4690]: E0217 00:06:34.977531 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.009786 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:35 crc kubenswrapper[4690]: E0217 00:06:35.009949 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:35 crc kubenswrapper[4690]: E0217 00:06:35.010041 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs podName:9050c3d5-4d74-4b57-afba-1dd177ce7983 nodeName:}" failed. No retries permitted until 2026-02-17 00:07:07.010014595 +0000 UTC m=+100.745727386 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs") pod "network-metrics-daemon-lz646" (UID: "9050c3d5-4d74-4b57-afba-1dd177ce7983") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.051375 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.051427 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.051439 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.051454 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.051466 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:35Z","lastTransitionTime":"2026-02-17T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.153577 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.153630 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.153640 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.153654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.153665 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:35Z","lastTransitionTime":"2026-02-17T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.256216 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.256285 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.256303 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.256716 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.256771 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:35Z","lastTransitionTime":"2026-02-17T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.359197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.359258 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.359281 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.359304 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.359321 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:35Z","lastTransitionTime":"2026-02-17T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.463343 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.463474 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.463497 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.463530 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.463554 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:35Z","lastTransitionTime":"2026-02-17T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.566122 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.566488 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.566496 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.566510 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.566518 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:35Z","lastTransitionTime":"2026-02-17T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.669086 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.669150 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.669171 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.669199 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.669218 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:35Z","lastTransitionTime":"2026-02-17T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.772180 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.772242 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.772255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.772279 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.772291 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:35Z","lastTransitionTime":"2026-02-17T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.875383 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.875432 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.875444 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.875462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.875474 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:35Z","lastTransitionTime":"2026-02-17T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.949455 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 06:13:52.192498788 +0000 UTC Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.977100 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.977144 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.977157 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.977171 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:35 crc kubenswrapper[4690]: I0217 00:06:35.977182 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:35Z","lastTransitionTime":"2026-02-17T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.079550 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.079602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.079616 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.079632 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.079643 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:36Z","lastTransitionTime":"2026-02-17T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.182041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.182095 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.182106 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.182123 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.182134 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:36Z","lastTransitionTime":"2026-02-17T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.284898 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.284968 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.284985 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.285009 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.285027 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:36Z","lastTransitionTime":"2026-02-17T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.387145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.387182 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.387191 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.387204 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.387213 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:36Z","lastTransitionTime":"2026-02-17T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.471906 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/0.log" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.471960 4690 generic.go:334] "Generic (PLEG): container finished" podID="797b14df-f7d0-419b-95f8-f02f25409e66" containerID="727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff" exitCode=1 Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.471997 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vb4v5" event={"ID":"797b14df-f7d0-419b-95f8-f02f25409e66","Type":"ContainerDied","Data":"727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.472427 4690 scope.go:117] "RemoveContainer" containerID="727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.489734 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.489765 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.489773 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.489786 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.489797 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:36Z","lastTransitionTime":"2026-02-17T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.492875 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:35Z\\\",\\\"message\\\":\\\"2026-02-17T00:05:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40\\\\n2026-02-17T00:05:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40 to /host/opt/cni/bin/\\\\n2026-02-17T00:05:50Z [verbose] multus-daemon started\\\\n2026-02-17T00:05:50Z [verbose] Readiness Indicator file check\\\\n2026-02-17T00:06:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.513559 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.529910 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.546101 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.566153 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.578302 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8057003-037a-43f0-a29f-8b30d2f9f973\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://971fe3f679431845df0be1a451c36f170cd038649d687a730a6b74adfe370b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dae842d1ef134d705b87402029edb7e1acc4d3b1efbc458eb8ceaf07f2426db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d468f64b7ca19373b20612c77a5482144dffdcd3e76d0377af6a421a4770d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.593104 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.593173 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.593189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.593217 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.593234 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:36Z","lastTransitionTime":"2026-02-17T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.594907 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.611111 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.622598 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.642146 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.673448 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"1 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:7FOCZ3GVMQ1pwQKJahWmE09uJDRx6ab8xxcEYE] map[] [{operators.coreos.com/v1alpha1 CatalogSource certified-operators faf75094-01cc-4ebf-8761-1361fa2bf31e 0xc0076f7a2d 0xc0076f7a2e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: certified-operators,olm.managed: true,},ClusterIP:10.217.5.214,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.214],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0217 00:06:12.996747 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.696056 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.696121 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.696135 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.696062 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.696156 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.696304 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:36Z","lastTransitionTime":"2026-02-17T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.707698 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.718662 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.734296 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.749661 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.763316 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.798593 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.798632 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.798678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.798702 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.798720 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:36Z","lastTransitionTime":"2026-02-17T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.902272 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.902323 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.902342 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.902399 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.902421 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:36Z","lastTransitionTime":"2026-02-17T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.951051 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 23:05:32.88337261 +0000 UTC Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.975558 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.975605 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.975663 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:36 crc kubenswrapper[4690]: E0217 00:06:36.975741 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.975826 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:36 crc kubenswrapper[4690]: E0217 00:06:36.975990 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:36 crc kubenswrapper[4690]: E0217 00:06:36.976046 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:36 crc kubenswrapper[4690]: E0217 00:06:36.976146 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:36 crc kubenswrapper[4690]: I0217 00:06:36.997260 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"1 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:7FOCZ3GVMQ1pwQKJahWmE09uJDRx6ab8xxcEYE] map[] [{operators.coreos.com/v1alpha1 CatalogSource certified-operators faf75094-01cc-4ebf-8761-1361fa2bf31e 0xc0076f7a2d 0xc0076f7a2e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: certified-operators,olm.managed: true,},ClusterIP:10.217.5.214,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.214],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0217 00:06:12.996747 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.006940 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.006988 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.007003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.007023 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.007038 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:37Z","lastTransitionTime":"2026-02-17T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.018096 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.050429 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.068432 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8057003-037a-43f0-a29f-8b30d2f9f973\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://971fe3f679431845df0be1a451c36f170cd038649d687a730a6b74adfe370b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dae842d1ef134d705b87402029edb7e1acc4d3b1efbc458eb8ceaf07f2426db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d468f64b7ca19373b20612c77a5482144dffdcd3e76d0377af6a421a4770d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.096071 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.110093 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.110146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.110156 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.110175 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.110187 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:37Z","lastTransitionTime":"2026-02-17T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.110344 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.120606 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.133621 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.146752 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.157283 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.172544 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.187250 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.199288 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.213427 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:35Z\\\",\\\"message\\\":\\\"2026-02-17T00:05:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40\\\\n2026-02-17T00:05:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40 to /host/opt/cni/bin/\\\\n2026-02-17T00:05:50Z [verbose] multus-daemon started\\\\n2026-02-17T00:05:50Z [verbose] Readiness Indicator file check\\\\n2026-02-17T00:06:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.213539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.213588 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.213602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.213625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.213641 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:37Z","lastTransitionTime":"2026-02-17T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.228516 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.243084 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.252964 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.316967 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.317008 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.317020 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.317035 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.317047 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:37Z","lastTransitionTime":"2026-02-17T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.419798 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.419841 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.419853 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.419868 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.419879 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:37Z","lastTransitionTime":"2026-02-17T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.475986 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/0.log" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.476048 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vb4v5" event={"ID":"797b14df-f7d0-419b-95f8-f02f25409e66","Type":"ContainerStarted","Data":"543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.489066 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.502783 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.512916 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.522030 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.522177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.522246 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.522335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.522444 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:37Z","lastTransitionTime":"2026-02-17T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.532538 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:35Z\\\",\\\"message\\\":\\\"2026-02-17T00:05:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40\\\\n2026-02-17T00:05:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40 to /host/opt/cni/bin/\\\\n2026-02-17T00:05:50Z [verbose] multus-daemon started\\\\n2026-02-17T00:05:50Z [verbose] Readiness Indicator file check\\\\n2026-02-17T00:06:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.547952 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.561243 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.573327 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.585259 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.597290 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.617695 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"1 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:7FOCZ3GVMQ1pwQKJahWmE09uJDRx6ab8xxcEYE] map[] [{operators.coreos.com/v1alpha1 CatalogSource certified-operators faf75094-01cc-4ebf-8761-1361fa2bf31e 0xc0076f7a2d 0xc0076f7a2e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: certified-operators,olm.managed: true,},ClusterIP:10.217.5.214,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.214],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0217 00:06:12.996747 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.624568 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.624593 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.624602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.624615 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.624624 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:37Z","lastTransitionTime":"2026-02-17T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.633075 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.652281 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.666249 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8057003-037a-43f0-a29f-8b30d2f9f973\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://971fe3f679431845df0be1a451c36f170cd038649d687a730a6b74adfe370b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dae842d1ef134d705b87402029edb7e1acc4d3b1efbc458eb8ceaf07f2426db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d468f64b7ca19373b20612c77a5482144dffdcd3e76d0377af6a421a4770d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.678649 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.691050 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.703480 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.715771 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.726490 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.726533 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.726543 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.726559 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.726568 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:37Z","lastTransitionTime":"2026-02-17T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.828927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.828951 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.828959 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.828971 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.828980 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:37Z","lastTransitionTime":"2026-02-17T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.931447 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.931492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.931501 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.931513 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.931523 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:37Z","lastTransitionTime":"2026-02-17T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:37 crc kubenswrapper[4690]: I0217 00:06:37.951791 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 11:11:20.323442655 +0000 UTC Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.033400 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.033436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.033446 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.033461 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.033473 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:38Z","lastTransitionTime":"2026-02-17T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.135713 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.135754 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.135769 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.135784 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.135795 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:38Z","lastTransitionTime":"2026-02-17T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.237956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.238001 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.238013 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.238026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.238036 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:38Z","lastTransitionTime":"2026-02-17T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.340827 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.340862 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.340870 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.340884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.340893 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:38Z","lastTransitionTime":"2026-02-17T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.443400 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.443465 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.443476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.443492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.443504 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:38Z","lastTransitionTime":"2026-02-17T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.546301 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.546355 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.546388 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.546405 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.546416 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:38Z","lastTransitionTime":"2026-02-17T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.649547 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.649612 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.649631 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.649656 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.649672 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:38Z","lastTransitionTime":"2026-02-17T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.752237 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.752293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.752309 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.752333 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.752350 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:38Z","lastTransitionTime":"2026-02-17T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.855038 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.855076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.855087 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.855102 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.855113 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:38Z","lastTransitionTime":"2026-02-17T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.952034 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 22:21:43.768089309 +0000 UTC Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.958258 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.958295 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.958306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.958325 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.958338 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:38Z","lastTransitionTime":"2026-02-17T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.975532 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.975546 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.975639 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:38 crc kubenswrapper[4690]: E0217 00:06:38.975826 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:38 crc kubenswrapper[4690]: I0217 00:06:38.976125 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:38 crc kubenswrapper[4690]: E0217 00:06:38.976271 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:38 crc kubenswrapper[4690]: E0217 00:06:38.976624 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:38 crc kubenswrapper[4690]: E0217 00:06:38.976941 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.061477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.061518 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.061529 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.061544 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.061553 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:39Z","lastTransitionTime":"2026-02-17T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.164489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.164529 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.164537 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.164552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.164561 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:39Z","lastTransitionTime":"2026-02-17T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.268224 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.268282 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.268300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.268324 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.268343 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:39Z","lastTransitionTime":"2026-02-17T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.371154 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.371214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.371232 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.371264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.371284 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:39Z","lastTransitionTime":"2026-02-17T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.473887 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.473938 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.473952 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.473971 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.473987 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:39Z","lastTransitionTime":"2026-02-17T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.576842 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.576930 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.576947 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.576973 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.576993 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:39Z","lastTransitionTime":"2026-02-17T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.679299 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.679340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.679349 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.679374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.679384 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:39Z","lastTransitionTime":"2026-02-17T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.781163 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.781202 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.781212 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.781228 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.781238 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:39Z","lastTransitionTime":"2026-02-17T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.883928 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.883990 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.884011 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.884037 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.884106 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:39Z","lastTransitionTime":"2026-02-17T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.952793 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:23:18.442235386 +0000 UTC Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.985594 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.985627 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.985638 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.985654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:39 crc kubenswrapper[4690]: I0217 00:06:39.985665 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:39Z","lastTransitionTime":"2026-02-17T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.087939 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.087991 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.088008 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.088030 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.088047 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:40Z","lastTransitionTime":"2026-02-17T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.191225 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.191278 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.191290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.191306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.191318 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:40Z","lastTransitionTime":"2026-02-17T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.294174 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.294232 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.294254 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.294295 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.294319 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:40Z","lastTransitionTime":"2026-02-17T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.396909 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.396954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.396966 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.396984 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.396998 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:40Z","lastTransitionTime":"2026-02-17T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.499145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.499217 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.499235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.499263 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.499281 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:40Z","lastTransitionTime":"2026-02-17T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.602152 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.602200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.602211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.602230 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.602240 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:40Z","lastTransitionTime":"2026-02-17T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.705492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.705535 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.705547 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.705568 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.705582 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:40Z","lastTransitionTime":"2026-02-17T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.807963 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.808007 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.808019 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.808036 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.808047 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:40Z","lastTransitionTime":"2026-02-17T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.911447 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.911514 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.911532 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.911554 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.911571 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:40Z","lastTransitionTime":"2026-02-17T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.953180 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 21:58:44.343582006 +0000 UTC Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.975851 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.975957 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:40 crc kubenswrapper[4690]: E0217 00:06:40.975974 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.976107 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.976470 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:40 crc kubenswrapper[4690]: E0217 00:06:40.977024 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:40 crc kubenswrapper[4690]: E0217 00:06:40.977138 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.977069 4690 scope.go:117] "RemoveContainer" containerID="230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d" Feb 17 00:06:40 crc kubenswrapper[4690]: E0217 00:06:40.977256 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:40 crc kubenswrapper[4690]: I0217 00:06:40.989312 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.013787 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.013826 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.013840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.013856 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.013867 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:41Z","lastTransitionTime":"2026-02-17T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.116790 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.116856 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.116875 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.116901 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.116918 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:41Z","lastTransitionTime":"2026-02-17T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.220184 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.220239 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.220255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.220275 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.220289 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:41Z","lastTransitionTime":"2026-02-17T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.323701 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.323753 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.323770 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.323794 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.323811 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:41Z","lastTransitionTime":"2026-02-17T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.426815 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.426880 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.426892 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.426912 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.426925 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:41Z","lastTransitionTime":"2026-02-17T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.512874 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/2.log" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.515305 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.515781 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.529078 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.529120 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.529130 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.529147 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.529158 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:41Z","lastTransitionTime":"2026-02-17T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.535697 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.551795 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55a8e072-f33b-48f3-837a-0528f63b8e8c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa8689631e853adf44aaf69172bc5b39bb6a3246bc8c668516ce10a57e1d81b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.563466 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.572465 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.585534 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.597066 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8057003-037a-43f0-a29f-8b30d2f9f973\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://971fe3f679431845df0be1a451c36f170cd038649d687a730a6b74adfe370b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dae842d1ef134d705b87402029edb7e1acc4d3b1efbc458eb8ceaf07f2426db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d468f64b7ca19373b20612c77a5482144dffdcd3e76d0377af6a421a4770d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.610374 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.622275 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.631350 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.631430 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.631475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.631493 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.631516 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.631536 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:41Z","lastTransitionTime":"2026-02-17T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.646733 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.665049 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"1 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:7FOCZ3GVMQ1pwQKJahWmE09uJDRx6ab8xxcEYE] map[] [{operators.coreos.com/v1alpha1 CatalogSource certified-operators faf75094-01cc-4ebf-8761-1361fa2bf31e 0xc0076f7a2d 0xc0076f7a2e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: certified-operators,olm.managed: true,},ClusterIP:10.217.5.214,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.214],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0217 00:06:12.996747 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.679229 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.689240 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.700586 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.714757 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.726928 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.733567 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.733594 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.733602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.733615 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.733624 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:41Z","lastTransitionTime":"2026-02-17T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.738836 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.750333 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:35Z\\\",\\\"message\\\":\\\"2026-02-17T00:05:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40\\\\n2026-02-17T00:05:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40 to /host/opt/cni/bin/\\\\n2026-02-17T00:05:50Z [verbose] multus-daemon started\\\\n2026-02-17T00:05:50Z [verbose] Readiness Indicator file check\\\\n2026-02-17T00:06:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.836304 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.836345 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.836399 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.836422 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.836437 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:41Z","lastTransitionTime":"2026-02-17T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.938891 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.938933 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.938944 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.938960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.938971 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:41Z","lastTransitionTime":"2026-02-17T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:41 crc kubenswrapper[4690]: I0217 00:06:41.953663 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 17:37:16.017197525 +0000 UTC Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.042193 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.042241 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.042258 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.042282 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.042300 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:42Z","lastTransitionTime":"2026-02-17T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.146430 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.146496 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.146514 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.146542 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.146563 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:42Z","lastTransitionTime":"2026-02-17T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.250072 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.250145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.250170 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.250199 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.250226 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:42Z","lastTransitionTime":"2026-02-17T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.352752 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.352812 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.352831 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.352856 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.352876 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:42Z","lastTransitionTime":"2026-02-17T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.456079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.456131 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.456155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.456185 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.456207 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:42Z","lastTransitionTime":"2026-02-17T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.521335 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/3.log" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.522555 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/2.log" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.525861 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" exitCode=1 Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.525906 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340"} Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.525944 4690 scope.go:117] "RemoveContainer" containerID="230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.527190 4690 scope.go:117] "RemoveContainer" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:06:42 crc kubenswrapper[4690]: E0217 00:06:42.527480 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.551558 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.561982 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.562027 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.562051 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.562071 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.562083 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:42Z","lastTransitionTime":"2026-02-17T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.565579 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.583319 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.608170 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://230857197b18bb6ef0e210fec43925d221ac39c8192ca91845612fae5efb997d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:13Z\\\",\\\"message\\\":\\\"1 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:7FOCZ3GVMQ1pwQKJahWmE09uJDRx6ab8xxcEYE] map[] [{operators.coreos.com/v1alpha1 CatalogSource certified-operators faf75094-01cc-4ebf-8761-1361fa2bf31e 0xc0076f7a2d 0xc0076f7a2e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: certified-operators,olm.managed: true,},ClusterIP:10.217.5.214,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.214],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0217 00:06:12.996747 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:41Z\\\",\\\"message\\\":\\\"shift-machine-api/machine-api-controllers_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.167:8441: 10.217.4.167:8442: 10.217.4.167:8444:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {62af83f3-e0c8-4632-aaaa-17488566a9d8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0217 00:06:41.915217 6744 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z]\\\\nI0217 00:06:41.915200 6744 model\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.632764 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.652463 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.664551 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.664605 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.664623 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.664647 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.664665 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:42Z","lastTransitionTime":"2026-02-17T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.666043 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8057003-037a-43f0-a29f-8b30d2f9f973\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://971fe3f679431845df0be1a451c36f170cd038649d687a730a6b74adfe370b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dae842d1ef134d705b87402029edb7e1acc4d3b1efbc458eb8ceaf07f2426db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d468f64b7ca19373b20612c77a5482144dffdcd3e76d0377af6a421a4770d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.684721 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.697851 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.712860 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.730351 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.746959 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.762633 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.766803 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.766840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.766851 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.766866 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.766878 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:42Z","lastTransitionTime":"2026-02-17T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.780401 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:35Z\\\",\\\"message\\\":\\\"2026-02-17T00:05:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40\\\\n2026-02-17T00:05:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40 to /host/opt/cni/bin/\\\\n2026-02-17T00:05:50Z [verbose] multus-daemon started\\\\n2026-02-17T00:05:50Z [verbose] Readiness Indicator file check\\\\n2026-02-17T00:06:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.795257 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.809132 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.824752 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55a8e072-f33b-48f3-837a-0528f63b8e8c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa8689631e853adf44aaf69172bc5b39bb6a3246bc8c668516ce10a57e1d81b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.837915 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:42Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.874212 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.874276 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.874293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.874317 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.874334 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:42Z","lastTransitionTime":"2026-02-17T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.953916 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 15:06:10.692140629 +0000 UTC Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.975299 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.975385 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:42 crc kubenswrapper[4690]: E0217 00:06:42.975486 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.975504 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:42 crc kubenswrapper[4690]: E0217 00:06:42.975649 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.975759 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:42 crc kubenswrapper[4690]: E0217 00:06:42.975938 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:42 crc kubenswrapper[4690]: E0217 00:06:42.976103 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.976592 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.976627 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.976643 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.976662 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:42 crc kubenswrapper[4690]: I0217 00:06:42.976673 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:42Z","lastTransitionTime":"2026-02-17T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.079839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.079916 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.079939 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.079971 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.079992 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:43Z","lastTransitionTime":"2026-02-17T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.182822 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.182918 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.182936 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.182966 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.182989 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:43Z","lastTransitionTime":"2026-02-17T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.285814 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.285884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.285909 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.285939 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.285960 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:43Z","lastTransitionTime":"2026-02-17T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.388953 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.389015 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.389026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.389046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.389058 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:43Z","lastTransitionTime":"2026-02-17T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.492065 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.492127 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.492150 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.492178 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.492199 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:43Z","lastTransitionTime":"2026-02-17T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.533738 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/3.log" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.539925 4690 scope.go:117] "RemoveContainer" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:06:43 crc kubenswrapper[4690]: E0217 00:06:43.540215 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.564269 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.588842 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.595148 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.595201 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.595219 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.595245 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.595265 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:43Z","lastTransitionTime":"2026-02-17T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.607062 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8057003-037a-43f0-a29f-8b30d2f9f973\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://971fe3f679431845df0be1a451c36f170cd038649d687a730a6b74adfe370b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dae842d1ef134d705b87402029edb7e1acc4d3b1efbc458eb8ceaf07f2426db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d468f64b7ca19373b20612c77a5482144dffdcd3e76d0377af6a421a4770d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.625053 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.643223 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.657669 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.675326 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.698614 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.698681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.698705 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.698735 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.698761 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:43Z","lastTransitionTime":"2026-02-17T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.700667 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:41Z\\\",\\\"message\\\":\\\"shift-machine-api/machine-api-controllers_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.167:8441: 10.217.4.167:8442: 10.217.4.167:8444:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {62af83f3-e0c8-4632-aaaa-17488566a9d8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0217 00:06:41.915217 6744 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z]\\\\nI0217 00:06:41.915200 6744 model\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.716143 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.730811 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.750563 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.764475 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.783023 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.801566 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.801666 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.801693 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.801725 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.801747 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:43Z","lastTransitionTime":"2026-02-17T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.804863 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:35Z\\\",\\\"message\\\":\\\"2026-02-17T00:05:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40\\\\n2026-02-17T00:05:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40 to /host/opt/cni/bin/\\\\n2026-02-17T00:05:50Z [verbose] multus-daemon started\\\\n2026-02-17T00:05:50Z [verbose] Readiness Indicator file check\\\\n2026-02-17T00:06:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.824323 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.841002 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55a8e072-f33b-48f3-837a-0528f63b8e8c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa8689631e853adf44aaf69172bc5b39bb6a3246bc8c668516ce10a57e1d81b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.858824 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.875764 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:43Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.904325 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.904456 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.904484 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.904515 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.904536 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:43Z","lastTransitionTime":"2026-02-17T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:43 crc kubenswrapper[4690]: I0217 00:06:43.954664 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 10:02:25.957883846 +0000 UTC Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.007426 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.007481 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.007502 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.007525 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.007541 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.110215 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.110255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.110268 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.110287 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.110304 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.212769 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.213190 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.213404 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.213725 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.213949 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.239707 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.239788 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.239806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.239832 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.239851 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: E0217 00:06:44.260086 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:44Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.265986 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.266046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.266067 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.266103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.266122 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: E0217 00:06:44.290819 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:44Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.295103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.295224 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.295294 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.295319 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.295335 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: E0217 00:06:44.316591 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:44Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.321729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.321797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.321820 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.321845 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.321861 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: E0217 00:06:44.341074 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:44Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.345618 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.345695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.345715 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.345746 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.345765 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: E0217 00:06:44.362749 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:44Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:44 crc kubenswrapper[4690]: E0217 00:06:44.363074 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.365190 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.365269 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.365293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.365336 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.365401 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.468689 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.469024 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.469221 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.469513 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.469738 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.572675 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.572751 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.572777 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.572805 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.572829 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.676133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.676202 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.676228 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.676260 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.676283 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.779543 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.779619 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.779681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.779713 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.779735 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.882158 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.882205 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.882219 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.882236 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.882249 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.955980 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 13:32:40.708463112 +0000 UTC Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.975693 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.975764 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.975822 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:44 crc kubenswrapper[4690]: E0217 00:06:44.975912 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:44 crc kubenswrapper[4690]: E0217 00:06:44.976172 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:44 crc kubenswrapper[4690]: E0217 00:06:44.976292 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.976625 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:44 crc kubenswrapper[4690]: E0217 00:06:44.976821 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.984677 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.984730 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.984748 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.984771 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:44 crc kubenswrapper[4690]: I0217 00:06:44.984790 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:44Z","lastTransitionTime":"2026-02-17T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.087959 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.088014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.088030 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.088054 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.088070 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:45Z","lastTransitionTime":"2026-02-17T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.191894 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.191931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.191941 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.191956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.191968 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:45Z","lastTransitionTime":"2026-02-17T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.315467 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.315537 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.315559 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.315584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.315602 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:45Z","lastTransitionTime":"2026-02-17T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.419297 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.419449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.419476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.419508 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.419530 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:45Z","lastTransitionTime":"2026-02-17T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.523238 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.523305 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.523323 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.523349 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.523396 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:45Z","lastTransitionTime":"2026-02-17T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.626492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.626560 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.626579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.626606 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.626623 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:45Z","lastTransitionTime":"2026-02-17T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.730084 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.730146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.730159 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.730183 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.730203 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:45Z","lastTransitionTime":"2026-02-17T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.833569 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.833669 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.833686 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.833709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.833722 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:45Z","lastTransitionTime":"2026-02-17T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.936463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.936531 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.936549 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.936582 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.936602 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:45Z","lastTransitionTime":"2026-02-17T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:45 crc kubenswrapper[4690]: I0217 00:06:45.956607 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 17:12:50.60360173 +0000 UTC Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.038943 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.038998 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.039014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.039040 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.039058 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:46Z","lastTransitionTime":"2026-02-17T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.141959 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.141998 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.142010 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.142025 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.142034 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:46Z","lastTransitionTime":"2026-02-17T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.244988 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.245073 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.245096 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.245126 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.245147 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:46Z","lastTransitionTime":"2026-02-17T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.348044 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.348103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.348119 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.348144 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.348159 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:46Z","lastTransitionTime":"2026-02-17T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.451153 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.451213 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.451235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.451264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.451287 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:46Z","lastTransitionTime":"2026-02-17T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.553692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.553732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.553745 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.553765 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.553780 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:46Z","lastTransitionTime":"2026-02-17T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.656713 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.656789 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.656814 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.656843 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.656861 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:46Z","lastTransitionTime":"2026-02-17T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.759975 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.760028 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.760044 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.760064 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.760080 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:46Z","lastTransitionTime":"2026-02-17T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.863515 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.863601 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.863618 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.863636 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.863651 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:46Z","lastTransitionTime":"2026-02-17T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.957629 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 11:37:46.724090645 +0000 UTC Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.966861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.966914 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.966938 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.966969 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.967026 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:46Z","lastTransitionTime":"2026-02-17T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.975309 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.975319 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.975479 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:46 crc kubenswrapper[4690]: E0217 00:06:46.975659 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.975761 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:46 crc kubenswrapper[4690]: E0217 00:06:46.975884 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:46 crc kubenswrapper[4690]: E0217 00:06:46.975990 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:46 crc kubenswrapper[4690]: E0217 00:06:46.976085 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:46 crc kubenswrapper[4690]: I0217 00:06:46.997794 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:35Z\\\",\\\"message\\\":\\\"2026-02-17T00:05:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40\\\\n2026-02-17T00:05:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40 to /host/opt/cni/bin/\\\\n2026-02-17T00:05:50Z [verbose] multus-daemon started\\\\n2026-02-17T00:05:50Z [verbose] Readiness Indicator file check\\\\n2026-02-17T00:06:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:46Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.019099 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.037730 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55a8e072-f33b-48f3-837a-0528f63b8e8c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa8689631e853adf44aaf69172bc5b39bb6a3246bc8c668516ce10a57e1d81b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.056023 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.070774 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.070986 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.071065 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.071137 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.071177 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:47Z","lastTransitionTime":"2026-02-17T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.075093 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.107322 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:41Z\\\",\\\"message\\\":\\\"shift-machine-api/machine-api-controllers_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.167:8441: 10.217.4.167:8442: 10.217.4.167:8444:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {62af83f3-e0c8-4632-aaaa-17488566a9d8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0217 00:06:41.915217 6744 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z]\\\\nI0217 00:06:41.915200 6744 model\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.130624 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.151323 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.169192 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8057003-037a-43f0-a29f-8b30d2f9f973\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://971fe3f679431845df0be1a451c36f170cd038649d687a730a6b74adfe370b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dae842d1ef134d705b87402029edb7e1acc4d3b1efbc458eb8ceaf07f2426db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d468f64b7ca19373b20612c77a5482144dffdcd3e76d0377af6a421a4770d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.174337 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.174435 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.174463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.174491 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.174510 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:47Z","lastTransitionTime":"2026-02-17T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.190240 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.211687 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.227791 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.245249 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.261188 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.278098 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.278156 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.278174 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.278197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.278213 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:47Z","lastTransitionTime":"2026-02-17T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.278919 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.298973 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.320349 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.336843 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:47Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.381024 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.381067 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.381089 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.381107 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.381119 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:47Z","lastTransitionTime":"2026-02-17T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.484380 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.484413 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.484424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.484438 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.484448 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:47Z","lastTransitionTime":"2026-02-17T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.587076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.587137 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.587156 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.587179 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.587196 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:47Z","lastTransitionTime":"2026-02-17T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.690516 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.690576 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.690594 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.690616 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.690632 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:47Z","lastTransitionTime":"2026-02-17T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.793752 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.793836 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.793860 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.793893 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.793921 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:47Z","lastTransitionTime":"2026-02-17T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.896947 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.897023 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.897046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.897077 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.897098 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:47Z","lastTransitionTime":"2026-02-17T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.958104 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 16:06:42.461874741 +0000 UTC Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.999659 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.999706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.999716 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:47 crc kubenswrapper[4690]: I0217 00:06:47.999728 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:47.999737 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:47Z","lastTransitionTime":"2026-02-17T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.108754 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.108829 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.108854 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.109640 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.109719 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:48Z","lastTransitionTime":"2026-02-17T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.213220 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.213262 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.213273 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.213290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.213299 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:48Z","lastTransitionTime":"2026-02-17T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.315723 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.315791 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.315808 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.315832 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.315854 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:48Z","lastTransitionTime":"2026-02-17T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.418210 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.418271 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.418285 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.418300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.418312 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:48Z","lastTransitionTime":"2026-02-17T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.521418 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.521468 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.521485 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.521507 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.521524 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:48Z","lastTransitionTime":"2026-02-17T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.624052 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.624138 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.624156 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.624180 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.624201 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:48Z","lastTransitionTime":"2026-02-17T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.727211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.727271 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.727289 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.727314 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.727332 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:48Z","lastTransitionTime":"2026-02-17T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.830079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.830151 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.830189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.830210 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.830221 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:48Z","lastTransitionTime":"2026-02-17T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.932787 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.932864 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.932882 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.932906 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.932925 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:48Z","lastTransitionTime":"2026-02-17T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.959046 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 18:47:32.224076206 +0000 UTC Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.975949 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.976052 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.975980 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:48 crc kubenswrapper[4690]: E0217 00:06:48.976170 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:48 crc kubenswrapper[4690]: I0217 00:06:48.976194 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:48 crc kubenswrapper[4690]: E0217 00:06:48.976304 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:48 crc kubenswrapper[4690]: E0217 00:06:48.976501 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:48 crc kubenswrapper[4690]: E0217 00:06:48.976649 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.036563 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.036590 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.036601 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.036615 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.036625 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:49Z","lastTransitionTime":"2026-02-17T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.138877 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.138919 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.138934 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.138983 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.139001 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:49Z","lastTransitionTime":"2026-02-17T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.242256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.242331 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.242396 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.242426 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.242444 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:49Z","lastTransitionTime":"2026-02-17T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.346328 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.346455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.346480 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.346515 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.346537 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:49Z","lastTransitionTime":"2026-02-17T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.449771 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.449812 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.449823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.449838 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.449849 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:49Z","lastTransitionTime":"2026-02-17T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.552796 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.552845 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.552863 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.552887 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.552904 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:49Z","lastTransitionTime":"2026-02-17T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.656436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.656485 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.656502 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.656524 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.656542 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:49Z","lastTransitionTime":"2026-02-17T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.759786 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.759849 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.759866 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.759892 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.759909 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:49Z","lastTransitionTime":"2026-02-17T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.863661 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.863738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.863762 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.863791 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.863815 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:49Z","lastTransitionTime":"2026-02-17T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.960083 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 22:31:47.908102778 +0000 UTC Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.966521 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.966590 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.966614 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.966679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:49 crc kubenswrapper[4690]: I0217 00:06:49.966701 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:49Z","lastTransitionTime":"2026-02-17T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.069748 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.069787 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.069798 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.069814 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.069824 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:50Z","lastTransitionTime":"2026-02-17T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.173342 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.173431 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.173448 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.173475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.173492 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:50Z","lastTransitionTime":"2026-02-17T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.276427 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.276502 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.276523 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.276552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.276574 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:50Z","lastTransitionTime":"2026-02-17T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.384306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.384442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.384469 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.384891 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.385205 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:50Z","lastTransitionTime":"2026-02-17T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.487895 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.487953 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.487972 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.487999 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.488017 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:50Z","lastTransitionTime":"2026-02-17T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.590881 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.590945 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.590967 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.590993 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.591013 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:50Z","lastTransitionTime":"2026-02-17T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.693639 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.694088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.694266 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.694501 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.694697 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:50Z","lastTransitionTime":"2026-02-17T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.799335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.799657 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.799794 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.799934 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.800065 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:50Z","lastTransitionTime":"2026-02-17T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.903398 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.903450 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.903466 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.903492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.903509 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:50Z","lastTransitionTime":"2026-02-17T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.960763 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 15:03:00.55270312 +0000 UTC Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.975139 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.975444 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:50 crc kubenswrapper[4690]: E0217 00:06:50.975651 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.975681 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.975684 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:50 crc kubenswrapper[4690]: E0217 00:06:50.975883 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:50 crc kubenswrapper[4690]: E0217 00:06:50.975999 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:50 crc kubenswrapper[4690]: E0217 00:06:50.976432 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:50 crc kubenswrapper[4690]: I0217 00:06:50.993652 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.006049 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.006242 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.006475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.006684 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.006851 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:51Z","lastTransitionTime":"2026-02-17T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.091205 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.091447 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.09141002 +0000 UTC m=+148.827122801 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.091720 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.091813 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.091863 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.091944 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.092029 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.092006428 +0000 UTC m=+148.827719219 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.092046 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.092076 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.092095 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.092149 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.092134491 +0000 UTC m=+148.827847282 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.092204 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.092243 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.092229494 +0000 UTC m=+148.827942285 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.110212 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.110269 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.110286 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.110312 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.110329 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:51Z","lastTransitionTime":"2026-02-17T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.193147 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.193395 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.193441 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.193461 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:51 crc kubenswrapper[4690]: E0217 00:06:51.193548 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.193520656 +0000 UTC m=+148.929233447 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.213646 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.213744 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.213764 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.213788 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.213846 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:51Z","lastTransitionTime":"2026-02-17T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.316896 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.317182 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.317200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.317222 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.317239 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:51Z","lastTransitionTime":"2026-02-17T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.420043 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.420084 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.420095 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.420132 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.420142 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:51Z","lastTransitionTime":"2026-02-17T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.522542 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.522587 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.522604 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.522622 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.522635 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:51Z","lastTransitionTime":"2026-02-17T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.625264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.625340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.625353 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.625389 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.625401 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:51Z","lastTransitionTime":"2026-02-17T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.727589 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.727669 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.727692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.727722 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.727785 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:51Z","lastTransitionTime":"2026-02-17T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.830437 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.830507 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.830519 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.830538 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.830550 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:51Z","lastTransitionTime":"2026-02-17T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.933168 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.933233 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.933252 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.933277 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.933296 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:51Z","lastTransitionTime":"2026-02-17T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:51 crc kubenswrapper[4690]: I0217 00:06:51.961641 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 08:40:33.685364122 +0000 UTC Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.038797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.038854 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.038875 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.038914 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.038934 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:52Z","lastTransitionTime":"2026-02-17T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.142608 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.142657 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.142672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.142695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.142709 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:52Z","lastTransitionTime":"2026-02-17T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.246333 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.246438 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.246457 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.246484 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.246500 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:52Z","lastTransitionTime":"2026-02-17T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.350002 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.350054 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.350137 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.350158 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.350169 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:52Z","lastTransitionTime":"2026-02-17T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.453924 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.453974 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.453997 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.454027 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.454048 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:52Z","lastTransitionTime":"2026-02-17T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.556454 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.556489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.556501 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.556516 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.556528 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:52Z","lastTransitionTime":"2026-02-17T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.659072 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.659115 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.659126 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.659143 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.659154 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:52Z","lastTransitionTime":"2026-02-17T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.762031 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.762139 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.762161 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.762186 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.762206 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:52Z","lastTransitionTime":"2026-02-17T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.864823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.864888 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.864910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.864937 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.864956 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:52Z","lastTransitionTime":"2026-02-17T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.961791 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 18:55:22.364199638 +0000 UTC Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.968354 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.968414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.968424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.968442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.968457 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:52Z","lastTransitionTime":"2026-02-17T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.976161 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.976164 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.976383 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:52 crc kubenswrapper[4690]: I0217 00:06:52.976432 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:52 crc kubenswrapper[4690]: E0217 00:06:52.976625 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:52 crc kubenswrapper[4690]: E0217 00:06:52.976763 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:52 crc kubenswrapper[4690]: E0217 00:06:52.976868 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:52 crc kubenswrapper[4690]: E0217 00:06:52.976946 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.072076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.072128 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.072141 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.072161 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.072174 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:53Z","lastTransitionTime":"2026-02-17T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.175999 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.176069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.176087 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.176109 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.176124 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:53Z","lastTransitionTime":"2026-02-17T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.278633 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.278695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.278711 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.278735 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.278751 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:53Z","lastTransitionTime":"2026-02-17T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.381716 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.381773 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.381799 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.381830 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.381853 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:53Z","lastTransitionTime":"2026-02-17T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.485256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.485299 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.485323 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.485344 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.485383 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:53Z","lastTransitionTime":"2026-02-17T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.588103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.588152 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.588171 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.588192 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.588209 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:53Z","lastTransitionTime":"2026-02-17T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.691055 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.691117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.691142 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.691171 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.691192 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:53Z","lastTransitionTime":"2026-02-17T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.794627 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.794718 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.794735 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.794761 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.794777 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:53Z","lastTransitionTime":"2026-02-17T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.898672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.898768 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.898784 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.898809 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.898828 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:53Z","lastTransitionTime":"2026-02-17T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:53 crc kubenswrapper[4690]: I0217 00:06:53.961975 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 14:50:49.192335704 +0000 UTC Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.002049 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.002138 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.002162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.002196 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.002218 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.105716 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.105783 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.105796 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.105815 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.105832 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.209095 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.209172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.209197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.209228 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.209250 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.312958 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.313007 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.313024 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.313046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.313060 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.415751 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.415804 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.415822 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.415847 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.415865 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.524818 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.524890 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.524912 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.524940 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.524961 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.627999 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.628073 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.628099 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.628131 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.628155 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.729860 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.729965 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.729987 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.730007 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.730022 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: E0217 00:06:54.748002 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.751835 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.751873 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.751885 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.751907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.751922 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: E0217 00:06:54.766597 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.770070 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.770115 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.770128 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.770148 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.770157 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: E0217 00:06:54.784643 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.788140 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.788196 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.788211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.788228 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.788238 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: E0217 00:06:54.802624 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.805754 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.805816 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.805830 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.805853 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.805868 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: E0217 00:06:54.822114 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9d0fc7d-16eb-44b8-b6b2-b3166201c8d1\\\",\\\"systemUUID\\\":\\\"ed6e37f0-bc2a-4f72-8a62-fab92258d3ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:54 crc kubenswrapper[4690]: E0217 00:06:54.822362 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.823911 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.823946 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.823955 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.823982 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.823992 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.926765 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.926829 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.926848 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.926875 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.926892 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:54Z","lastTransitionTime":"2026-02-17T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.962139 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 00:58:35.497688921 +0000 UTC Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.975843 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.975878 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.975965 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:54 crc kubenswrapper[4690]: E0217 00:06:54.976145 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:54 crc kubenswrapper[4690]: I0217 00:06:54.976178 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:54 crc kubenswrapper[4690]: E0217 00:06:54.976315 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:54 crc kubenswrapper[4690]: E0217 00:06:54.976647 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:54 crc kubenswrapper[4690]: E0217 00:06:54.976715 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.030151 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.030211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.030231 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.030256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.030275 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:55Z","lastTransitionTime":"2026-02-17T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.132951 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.132981 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.132990 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.133003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.133012 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:55Z","lastTransitionTime":"2026-02-17T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.236893 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.236934 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.236943 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.236959 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.236967 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:55Z","lastTransitionTime":"2026-02-17T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.340915 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.341019 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.341072 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.341099 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.341118 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:55Z","lastTransitionTime":"2026-02-17T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.446126 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.446196 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.446217 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.446246 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.446266 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:55Z","lastTransitionTime":"2026-02-17T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.550721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.550786 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.550804 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.550828 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.550846 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:55Z","lastTransitionTime":"2026-02-17T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.654834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.654913 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.654937 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.654966 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.654987 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:55Z","lastTransitionTime":"2026-02-17T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.759085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.759201 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.759221 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.759250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.759268 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:55Z","lastTransitionTime":"2026-02-17T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.863739 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.863797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.863807 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.863825 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.863841 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:55Z","lastTransitionTime":"2026-02-17T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.963436 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 00:30:38.481325437 +0000 UTC Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.967311 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.967429 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.967459 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.967485 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:55 crc kubenswrapper[4690]: I0217 00:06:55.967504 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:55Z","lastTransitionTime":"2026-02-17T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.070083 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.070176 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.070184 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.070217 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.070229 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:56Z","lastTransitionTime":"2026-02-17T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.172872 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.172995 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.173009 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.173026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.173039 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:56Z","lastTransitionTime":"2026-02-17T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.276160 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.276209 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.276227 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.276250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.276266 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:56Z","lastTransitionTime":"2026-02-17T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.379317 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.379398 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.379417 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.379442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.379459 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:56Z","lastTransitionTime":"2026-02-17T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.482199 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.482260 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.482272 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.482293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.482306 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:56Z","lastTransitionTime":"2026-02-17T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.584801 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.584857 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.584874 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.584897 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.584914 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:56Z","lastTransitionTime":"2026-02-17T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.687873 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.687960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.687980 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.688011 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.688035 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:56Z","lastTransitionTime":"2026-02-17T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.791784 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.791846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.791863 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.791892 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.791909 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:56Z","lastTransitionTime":"2026-02-17T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.895287 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.895344 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.895406 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.895438 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.895460 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:56Z","lastTransitionTime":"2026-02-17T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.964342 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 11:48:47.590431046 +0000 UTC Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.975037 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.975136 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.975234 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.976109 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:56 crc kubenswrapper[4690]: E0217 00:06:56.976230 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:56 crc kubenswrapper[4690]: E0217 00:06:56.976583 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:56 crc kubenswrapper[4690]: E0217 00:06:56.977858 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:56 crc kubenswrapper[4690]: E0217 00:06:56.977869 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.978516 4690 scope.go:117] "RemoveContainer" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:06:56 crc kubenswrapper[4690]: E0217 00:06:56.978805 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.999578 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.999625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.999647 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:56 crc kubenswrapper[4690]: I0217 00:06:56.999673 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:56.999695 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:56Z","lastTransitionTime":"2026-02-17T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.009963 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5154e499-f217-4a23-82cf-695e89168863\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fa08526608473b932b6638485c2291690004e13a0e409a1dde28e531a89a166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e53310be1ede7135c52d6f9f9c3a996a42129a3bfa062bd0145038944da246ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85bae26ae90f236917810f277dedea3283d1ceb530696a3aa5036ce1cc1ac297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f4b80ca314b3a59b982376a4fb6c98a850a7ef5a2cd1117387e414d65b61d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2297b4caaf656677fc3d5011bd283d0e861e916f76e9e4e92e8ca12a4c211e9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3895b844cf90fffb974f0e328d15c64e1ce6bcf2e5cabc3997201315ca8289a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3895b844cf90fffb974f0e328d15c64e1ce6bcf2e5cabc3997201315ca8289a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced2a900f58d82b450f0f25d729563ff1fd783ce1ee74f94cbcdb752d235ae76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ced2a900f58d82b450f0f25d729563ff1fd783ce1ee74f94cbcdb752d235ae76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3f1b6e538bb6bf2ca276d357fae177c5a103541eb610a18cca2ecfce6078cae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1b6e538bb6bf2ca276d357fae177c5a103541eb610a18cca2ecfce6078cae7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.031922 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vb4v5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797b14df-f7d0-419b-95f8-f02f25409e66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:35Z\\\",\\\"message\\\":\\\"2026-02-17T00:05:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40\\\\n2026-02-17T00:05:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_903eee5d-1a3b-4b18-a305-a5069c9b0b40 to /host/opt/cni/bin/\\\\n2026-02-17T00:05:50Z [verbose] multus-daemon started\\\\n2026-02-17T00:05:50Z [verbose] Readiness Indicator file check\\\\n2026-02-17T00:06:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfv9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vb4v5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.055208 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bda0ad83-a9db-4baf-9df4-4cac4b9be7c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b6c6452450ca03321424f48be2f23dd5a2082d0873c22bdc5ce08ece289de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbffd2d0828e3bfc28ae18c1fab8d896ffbbfa43b45c4184ba054dec67c47c1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2f3b55f64ec620eb191c60d497ecdf482484e2ff526265ce970dee46d18f4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.074217 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55a8e072-f33b-48f3-837a-0528f63b8e8c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa8689631e853adf44aaf69172bc5b39bb6a3246bc8c668516ce10a57e1d81b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec4b77d3d98a704d42c70650638c3b0de46ea99fae6e24c558fe6979dea79e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.097163 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6fe888eda8332d9af88f77274117f5845830314acd93e21565f0914f8596bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.103901 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.103949 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.103967 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.103990 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.104007 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:57Z","lastTransitionTime":"2026-02-17T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.113678 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f04a05d6-a6e4-42ac-bb58-99be30f6a38a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aaa8a826cc9ac59d53f039f4f8841362f44a1f589f220687e34a49ce196782da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd504fabc0bb69d01f7eee41b022097bab7e1de251d5dcb34409850591d5d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvjlh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qw6ns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.144004 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T00:06:41Z\\\",\\\"message\\\":\\\"shift-machine-api/machine-api-controllers_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.167:8441: 10.217.4.167:8442: 10.217.4.167:8444:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {62af83f3-e0c8-4632-aaaa-17488566a9d8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0217 00:06:41.915217 6744 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:41Z is after 2025-08-24T17:21:41Z]\\\\nI0217 00:06:41.915200 6744 model\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2frwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.161927 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pw72j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5246f770-6fb0-4762-97c7-9473fcc26738\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21812b81d8d3f75deae85c749007b0af5c7e32ceb0eed546097194ca241deb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fef4801c47a82bb8d9ce740ae224ade121638e725ee6ac04f9f6d52ac7fc171\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21296c4ccb61fe87372d788507ab1d583b177b1d31cf447201edbede3ad305b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dd9b546bd09c8003d1b7a4170a308e67f9b2ea7a0663d0b2ce6e4a59e08e32e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59bcc2b1381bce543ad3e947dcada5ae9c7bc9e9122a97e0052feb1825be5e47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769b06222cc584f15bb1d5019c308a784a4eef91bed3bb705cb466d8600d8a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04030ac5892f78b0f7c0c7f392f0a3e9006b1ff321682eda0ed77a90d95487d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2lxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pw72j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.182510 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac3fd592-6a9c-44aa-87bd-31732029af22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T00:05:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 00:05:40.608973 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 00:05:40.612043 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3095007605/tls.crt::/tmp/serving-cert-3095007605/tls.key\\\\\\\"\\\\nI0217 00:05:46.648022 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 00:05:46.662793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 00:05:46.662831 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 00:05:46.662861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 00:05:46.662871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 00:05:46.672930 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 00:05:46.672973 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 00:05:46.672999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 00:05:46.673006 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 00:05:46.673015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 00:05:46.673023 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 00:05:46.673028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 00:05:46.676697 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.196081 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8057003-037a-43f0-a29f-8b30d2f9f973\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://971fe3f679431845df0be1a451c36f170cd038649d687a730a6b74adfe370b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dae842d1ef134d705b87402029edb7e1acc4d3b1efbc458eb8ceaf07f2426db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d468f64b7ca19373b20612c77a5482144dffdcd3e76d0377af6a421a4770d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efedbe47c66bd4956b01d8e010b0282cd63caee467768914a7afb725dbdeb251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T00:05:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T00:05:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.206412 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.206468 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.206484 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.206502 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.206514 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:57Z","lastTransitionTime":"2026-02-17T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.212197 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc5fe1f372dba72ac9268f94817462fe9611e6c0b9373aac7f172ebc614644d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.229923 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.246491 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v5cn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9cb4ddb-30f5-476e-b815-a847fe68e60f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6483148c28824765d6d62e1a536343f08ff2e4a95f22f66334cd6be526162408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-527vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v5cn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.260176 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8868c633-9a82-4998-b1b4-06c6d76396ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19968fd12d1193aa602af22d8610ae357d299f00ec7b1acc02dd09911dc4ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s6gtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lszwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.275870 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wpnhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea8d7ce4-2bc3-4db6-8a8e-467d9a5bac30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0c18068bf8fd2d54df8fb2420055a60fa826315109433396d97e43070cbdf9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpndk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:05:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wpnhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.294957 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lz646" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9050c3d5-4d74-4b57-afba-1dd177ce7983\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T00:06:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7982r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T00:06:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lz646\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.310210 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.310260 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.310270 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.310285 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.310294 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:57Z","lastTransitionTime":"2026-02-17T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.317314 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b1d7c7a86ab6e493ab58c1ee22c6dcec6ae4e91a2e095ff11733d35432081\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ccca3d73efe69de307b56749b6ed429880fa9f394c89b6312fdfe88c083390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T00:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.342321 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.362845 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T00:05:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.413898 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.413955 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.413965 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.413980 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.413988 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:57Z","lastTransitionTime":"2026-02-17T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.518834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.518908 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.518926 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.518958 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.518988 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:57Z","lastTransitionTime":"2026-02-17T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.623002 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.623062 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.623075 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.623094 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.623109 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:57Z","lastTransitionTime":"2026-02-17T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.726642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.726691 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.726701 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.726721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.726733 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:57Z","lastTransitionTime":"2026-02-17T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.830477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.830552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.830574 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.830604 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.830625 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:57Z","lastTransitionTime":"2026-02-17T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.933165 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.933235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.933260 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.933289 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.933315 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:57Z","lastTransitionTime":"2026-02-17T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:57 crc kubenswrapper[4690]: I0217 00:06:57.964591 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 22:44:17.40648322 +0000 UTC Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.036217 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.036276 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.036296 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.036319 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.036336 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:58Z","lastTransitionTime":"2026-02-17T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.138752 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.138816 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.138833 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.138859 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.138876 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:58Z","lastTransitionTime":"2026-02-17T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.242713 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.242779 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.242797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.242820 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.242838 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:58Z","lastTransitionTime":"2026-02-17T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.346111 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.346168 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.346188 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.346211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.346227 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:58Z","lastTransitionTime":"2026-02-17T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.449077 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.449190 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.449223 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.449251 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.449273 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:58Z","lastTransitionTime":"2026-02-17T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.552539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.552616 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.552652 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.552679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.552700 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:58Z","lastTransitionTime":"2026-02-17T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.656081 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.656146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.656177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.656220 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.656247 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:58Z","lastTransitionTime":"2026-02-17T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.759655 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.759711 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.759728 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.759755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.759777 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:58Z","lastTransitionTime":"2026-02-17T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.863021 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.863071 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.863084 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.863100 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.863111 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:58Z","lastTransitionTime":"2026-02-17T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.964730 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 00:39:08.672495098 +0000 UTC Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.967041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.967090 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.967101 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.967120 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.967133 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:58Z","lastTransitionTime":"2026-02-17T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.975652 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.975707 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.975709 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:06:58 crc kubenswrapper[4690]: E0217 00:06:58.975835 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:06:58 crc kubenswrapper[4690]: I0217 00:06:58.975767 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:06:58 crc kubenswrapper[4690]: E0217 00:06:58.975752 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:06:58 crc kubenswrapper[4690]: E0217 00:06:58.975998 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:06:58 crc kubenswrapper[4690]: E0217 00:06:58.976323 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.070636 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.070708 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.070729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.070760 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.070781 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:59Z","lastTransitionTime":"2026-02-17T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.180578 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.180740 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.180759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.180787 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.180817 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:59Z","lastTransitionTime":"2026-02-17T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.283524 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.283595 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.283607 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.283623 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.283634 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:59Z","lastTransitionTime":"2026-02-17T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.386397 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.386634 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.386715 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.386806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.386886 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:59Z","lastTransitionTime":"2026-02-17T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.489937 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.489996 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.490015 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.490039 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.490062 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:59Z","lastTransitionTime":"2026-02-17T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.593830 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.593939 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.593969 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.594001 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.594027 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:59Z","lastTransitionTime":"2026-02-17T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.697542 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.697643 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.697681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.697721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.697747 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:59Z","lastTransitionTime":"2026-02-17T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.801348 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.801470 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.801490 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.801513 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.801531 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:59Z","lastTransitionTime":"2026-02-17T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.903994 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.904069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.904087 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.904117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.904138 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:06:59Z","lastTransitionTime":"2026-02-17T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:06:59 crc kubenswrapper[4690]: I0217 00:06:59.965099 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 12:51:29.152122499 +0000 UTC Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.007543 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.007616 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.007643 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.007679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.007704 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:00Z","lastTransitionTime":"2026-02-17T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.109721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.109777 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.109815 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.109839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.109855 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:00Z","lastTransitionTime":"2026-02-17T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.212785 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.212883 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.212901 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.212930 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.212951 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:00Z","lastTransitionTime":"2026-02-17T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.316053 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.316109 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.316135 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.316164 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.316184 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:00Z","lastTransitionTime":"2026-02-17T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.419339 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.419462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.419489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.419519 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.419543 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:00Z","lastTransitionTime":"2026-02-17T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.521770 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.521810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.521820 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.521837 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.521850 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:00Z","lastTransitionTime":"2026-02-17T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.626063 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.626136 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.626153 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.626180 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.626198 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:00Z","lastTransitionTime":"2026-02-17T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.729872 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.729954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.729979 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.730010 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.730035 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:00Z","lastTransitionTime":"2026-02-17T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.833785 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.833832 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.833843 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.833861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.833876 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:00Z","lastTransitionTime":"2026-02-17T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.937234 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.937350 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.937432 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.937475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.937497 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:00Z","lastTransitionTime":"2026-02-17T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.966235 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 17:13:46.385807209 +0000 UTC Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.975877 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.975958 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.975961 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:00 crc kubenswrapper[4690]: E0217 00:07:00.976094 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:00 crc kubenswrapper[4690]: I0217 00:07:00.976236 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:00 crc kubenswrapper[4690]: E0217 00:07:00.976482 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:00 crc kubenswrapper[4690]: E0217 00:07:00.976627 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:00 crc kubenswrapper[4690]: E0217 00:07:00.976777 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.041318 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.041423 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.041449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.041479 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.041499 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:01Z","lastTransitionTime":"2026-02-17T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.144143 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.144244 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.144263 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.144287 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.144303 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:01Z","lastTransitionTime":"2026-02-17T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.248672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.248766 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.248790 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.248829 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.248851 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:01Z","lastTransitionTime":"2026-02-17T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.351996 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.352088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.352113 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.352145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.352172 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:01Z","lastTransitionTime":"2026-02-17T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.455878 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.455947 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.455957 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.455978 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.455990 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:01Z","lastTransitionTime":"2026-02-17T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.558287 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.558327 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.558336 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.558349 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.558387 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:01Z","lastTransitionTime":"2026-02-17T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.661983 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.662043 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.662060 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.662082 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.662099 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:01Z","lastTransitionTime":"2026-02-17T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.765189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.765235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.765247 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.765263 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.765274 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:01Z","lastTransitionTime":"2026-02-17T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.869250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.869332 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.869388 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.869422 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.869444 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:01Z","lastTransitionTime":"2026-02-17T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.966833 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 20:24:40.098878726 +0000 UTC Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.972300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.972417 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.972442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.972474 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:01 crc kubenswrapper[4690]: I0217 00:07:01.972498 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:01Z","lastTransitionTime":"2026-02-17T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.075476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.075526 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.075540 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.075558 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.075570 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:02Z","lastTransitionTime":"2026-02-17T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.178564 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.178659 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.178678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.178702 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.178720 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:02Z","lastTransitionTime":"2026-02-17T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.282310 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.282413 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.282432 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.282461 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.282479 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:02Z","lastTransitionTime":"2026-02-17T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.386799 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.386933 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.386956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.386983 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.387001 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:02Z","lastTransitionTime":"2026-02-17T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.490357 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.490496 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.490518 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.490539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.490556 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:02Z","lastTransitionTime":"2026-02-17T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.594955 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.595006 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.595017 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.595036 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.595048 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:02Z","lastTransitionTime":"2026-02-17T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.697948 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.698052 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.698079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.698120 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.698148 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:02Z","lastTransitionTime":"2026-02-17T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.801312 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.801443 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.801468 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.801502 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.801522 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:02Z","lastTransitionTime":"2026-02-17T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.905580 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.905642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.905660 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.905686 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.905704 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:02Z","lastTransitionTime":"2026-02-17T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.967543 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 20:22:39.540367331 +0000 UTC Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.975788 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.975892 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.975888 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:02 crc kubenswrapper[4690]: E0217 00:07:02.976020 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:02 crc kubenswrapper[4690]: I0217 00:07:02.976053 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:02 crc kubenswrapper[4690]: E0217 00:07:02.976151 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:02 crc kubenswrapper[4690]: E0217 00:07:02.976267 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:02 crc kubenswrapper[4690]: E0217 00:07:02.976392 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.009231 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.009280 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.009294 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.009322 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.009342 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:03Z","lastTransitionTime":"2026-02-17T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.112508 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.112599 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.112619 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.112645 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.112663 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:03Z","lastTransitionTime":"2026-02-17T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.215791 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.215850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.215868 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.215893 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.215911 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:03Z","lastTransitionTime":"2026-02-17T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.319047 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.319155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.319176 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.319237 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.319256 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:03Z","lastTransitionTime":"2026-02-17T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.422872 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.422950 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.422989 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.423018 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.423039 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:03Z","lastTransitionTime":"2026-02-17T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.526535 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.526654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.526671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.526697 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.526716 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:03Z","lastTransitionTime":"2026-02-17T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.629632 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.629705 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.629729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.629758 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.629780 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:03Z","lastTransitionTime":"2026-02-17T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.733242 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.733301 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.733325 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.733352 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.733420 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:03Z","lastTransitionTime":"2026-02-17T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.836505 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.836575 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.836592 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.836621 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.836636 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:03Z","lastTransitionTime":"2026-02-17T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.940180 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.940265 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.940306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.940341 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.940412 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:03Z","lastTransitionTime":"2026-02-17T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:03 crc kubenswrapper[4690]: I0217 00:07:03.968481 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 07:58:06.280427606 +0000 UTC Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.044591 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.044686 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.044711 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.044742 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.044763 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:04Z","lastTransitionTime":"2026-02-17T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.147785 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.147845 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.147862 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.147891 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.147908 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:04Z","lastTransitionTime":"2026-02-17T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.250806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.250860 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.250870 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.250911 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.250920 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:04Z","lastTransitionTime":"2026-02-17T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.353253 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.353305 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.353323 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.353348 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.353393 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:04Z","lastTransitionTime":"2026-02-17T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.460977 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.461014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.461024 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.461038 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.461047 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:04Z","lastTransitionTime":"2026-02-17T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.563759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.563819 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.563837 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.563861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.563878 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:04Z","lastTransitionTime":"2026-02-17T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.667499 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.667580 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.667598 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.667625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.667643 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:04Z","lastTransitionTime":"2026-02-17T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.776842 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.776902 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.776914 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.776929 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.776938 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:04Z","lastTransitionTime":"2026-02-17T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.880162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.880218 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.880230 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.880248 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.880260 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:04Z","lastTransitionTime":"2026-02-17T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.969678 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 03:19:39.368290835 +0000 UTC Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.975065 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.975111 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:04 crc kubenswrapper[4690]: E0217 00:07:04.975193 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.975214 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.975059 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:04 crc kubenswrapper[4690]: E0217 00:07:04.975420 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:04 crc kubenswrapper[4690]: E0217 00:07:04.975480 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:04 crc kubenswrapper[4690]: E0217 00:07:04.975542 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.982262 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.982328 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.982350 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.982414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:04 crc kubenswrapper[4690]: I0217 00:07:04.982438 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:04Z","lastTransitionTime":"2026-02-17T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.085425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.085540 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.085565 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.085596 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.085619 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:05Z","lastTransitionTime":"2026-02-17T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.183460 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.183533 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.183554 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.183583 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.183606 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T00:07:05Z","lastTransitionTime":"2026-02-17T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.253092 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9"] Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.253721 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.259247 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.259488 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.259530 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.260056 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.321075 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=15.321038965 podStartE2EDuration="15.321038965s" podCreationTimestamp="2026-02-17 00:06:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.297338576 +0000 UTC m=+99.033051397" watchObservedRunningTime="2026-02-17 00:07:05.321038965 +0000 UTC m=+99.056751726" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.323035 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-vb4v5" podStartSLOduration=78.323009412 podStartE2EDuration="1m18.323009412s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.317869193 +0000 UTC m=+99.053581954" watchObservedRunningTime="2026-02-17 00:07:05.323009412 +0000 UTC m=+99.058722203" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.337980 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=73.337928765 podStartE2EDuration="1m13.337928765s" podCreationTimestamp="2026-02-17 00:05:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.33706548 +0000 UTC m=+99.072778271" watchObservedRunningTime="2026-02-17 00:07:05.337928765 +0000 UTC m=+99.073641536" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.351027 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.351000845 podStartE2EDuration="25.351000845s" podCreationTimestamp="2026-02-17 00:06:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.351000035 +0000 UTC m=+99.086712786" watchObservedRunningTime="2026-02-17 00:07:05.351000845 +0000 UTC m=+99.086713626" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.382229 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qw6ns" podStartSLOduration=77.382206572 podStartE2EDuration="1m17.382206572s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.381492021 +0000 UTC m=+99.117204812" watchObservedRunningTime="2026-02-17 00:07:05.382206572 +0000 UTC m=+99.117919323" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.449914 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5b89ae6c-d4dd-46c1-9644-c3624ee34674-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.450045 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b89ae6c-d4dd-46c1-9644-c3624ee34674-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.450249 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5b89ae6c-d4dd-46c1-9644-c3624ee34674-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.450304 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b89ae6c-d4dd-46c1-9644-c3624ee34674-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.451205 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5b89ae6c-d4dd-46c1-9644-c3624ee34674-service-ca\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.470779 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-pw72j" podStartSLOduration=78.470761694 podStartE2EDuration="1m18.470761694s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.470054243 +0000 UTC m=+99.205766984" watchObservedRunningTime="2026-02-17 00:07:05.470761694 +0000 UTC m=+99.206474445" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.492782 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.492747733 podStartE2EDuration="1m18.492747733s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.49129264 +0000 UTC m=+99.227005431" watchObservedRunningTime="2026-02-17 00:07:05.492747733 +0000 UTC m=+99.228460484" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.521841 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=43.521814637 podStartE2EDuration="43.521814637s" podCreationTimestamp="2026-02-17 00:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.515681259 +0000 UTC m=+99.251394020" watchObservedRunningTime="2026-02-17 00:07:05.521814637 +0000 UTC m=+99.257527428" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.552092 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5b89ae6c-d4dd-46c1-9644-c3624ee34674-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.552144 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b89ae6c-d4dd-46c1-9644-c3624ee34674-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.552202 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5b89ae6c-d4dd-46c1-9644-c3624ee34674-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.552228 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b89ae6c-d4dd-46c1-9644-c3624ee34674-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.552246 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5b89ae6c-d4dd-46c1-9644-c3624ee34674-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.552285 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5b89ae6c-d4dd-46c1-9644-c3624ee34674-service-ca\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.552330 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5b89ae6c-d4dd-46c1-9644-c3624ee34674-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.553605 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5b89ae6c-d4dd-46c1-9644-c3624ee34674-service-ca\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.561116 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b89ae6c-d4dd-46c1-9644-c3624ee34674-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.573573 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b89ae6c-d4dd-46c1-9644-c3624ee34674-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-f9vn9\" (UID: \"5b89ae6c-d4dd-46c1-9644-c3624ee34674\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.580048 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.611840 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-v5cn7" podStartSLOduration=78.611812001 podStartE2EDuration="1m18.611812001s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.609889225 +0000 UTC m=+99.345601976" watchObservedRunningTime="2026-02-17 00:07:05.611812001 +0000 UTC m=+99.347524752" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.614470 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" event={"ID":"5b89ae6c-d4dd-46c1-9644-c3624ee34674","Type":"ContainerStarted","Data":"5b2e244a4d4d27efc2fbd1208c4192459ef0eb1a41ae80e81a239c1021c3feb5"} Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.640277 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podStartSLOduration=78.640254938 podStartE2EDuration="1m18.640254938s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.627853247 +0000 UTC m=+99.363565998" watchObservedRunningTime="2026-02-17 00:07:05.640254938 +0000 UTC m=+99.375967699" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.640668 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-wpnhj" podStartSLOduration=78.640661599 podStartE2EDuration="1m18.640661599s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:05.639505216 +0000 UTC m=+99.375217967" watchObservedRunningTime="2026-02-17 00:07:05.640661599 +0000 UTC m=+99.376374360" Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.970242 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 12:55:39.352501169 +0000 UTC Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.970426 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 17 00:07:05 crc kubenswrapper[4690]: I0217 00:07:05.982630 4690 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 17 00:07:06 crc kubenswrapper[4690]: I0217 00:07:06.619012 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" event={"ID":"5b89ae6c-d4dd-46c1-9644-c3624ee34674","Type":"ContainerStarted","Data":"ecede079eb55de1ffc6b5d99289d26889ac3725ca5dc39bc1fc6a25e669dbb05"} Feb 17 00:07:06 crc kubenswrapper[4690]: I0217 00:07:06.638649 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f9vn9" podStartSLOduration=79.638612338 podStartE2EDuration="1m19.638612338s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:06.638411392 +0000 UTC m=+100.374124143" watchObservedRunningTime="2026-02-17 00:07:06.638612338 +0000 UTC m=+100.374325129" Feb 17 00:07:06 crc kubenswrapper[4690]: I0217 00:07:06.976208 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:06 crc kubenswrapper[4690]: E0217 00:07:06.977490 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:06 crc kubenswrapper[4690]: I0217 00:07:06.977873 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:06 crc kubenswrapper[4690]: I0217 00:07:06.977931 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:06 crc kubenswrapper[4690]: I0217 00:07:06.978258 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:06 crc kubenswrapper[4690]: E0217 00:07:06.978896 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:06 crc kubenswrapper[4690]: E0217 00:07:06.979128 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:06 crc kubenswrapper[4690]: E0217 00:07:06.979345 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:07 crc kubenswrapper[4690]: E0217 00:07:07.065118 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:07:07 crc kubenswrapper[4690]: E0217 00:07:07.065268 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs podName:9050c3d5-4d74-4b57-afba-1dd177ce7983 nodeName:}" failed. No retries permitted until 2026-02-17 00:08:11.065229671 +0000 UTC m=+164.800942462 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs") pod "network-metrics-daemon-lz646" (UID: "9050c3d5-4d74-4b57-afba-1dd177ce7983") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 00:07:07 crc kubenswrapper[4690]: I0217 00:07:07.065586 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:08 crc kubenswrapper[4690]: I0217 00:07:08.975102 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:08 crc kubenswrapper[4690]: I0217 00:07:08.975175 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:08 crc kubenswrapper[4690]: E0217 00:07:08.980989 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:08 crc kubenswrapper[4690]: E0217 00:07:08.981172 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:08 crc kubenswrapper[4690]: I0217 00:07:08.981528 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:08 crc kubenswrapper[4690]: E0217 00:07:08.981861 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:08 crc kubenswrapper[4690]: I0217 00:07:08.980827 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:08 crc kubenswrapper[4690]: E0217 00:07:08.982215 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:10 crc kubenswrapper[4690]: I0217 00:07:10.975738 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:10 crc kubenswrapper[4690]: I0217 00:07:10.975776 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:10 crc kubenswrapper[4690]: E0217 00:07:10.975886 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:10 crc kubenswrapper[4690]: I0217 00:07:10.975943 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:10 crc kubenswrapper[4690]: I0217 00:07:10.976016 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:10 crc kubenswrapper[4690]: E0217 00:07:10.976174 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:10 crc kubenswrapper[4690]: E0217 00:07:10.976279 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:10 crc kubenswrapper[4690]: E0217 00:07:10.976342 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:11 crc kubenswrapper[4690]: I0217 00:07:11.977148 4690 scope.go:117] "RemoveContainer" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:07:11 crc kubenswrapper[4690]: E0217 00:07:11.977421 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2frwb_openshift-ovn-kubernetes(c6eeb20f-1ee2-4d96-a53b-da30d60213f3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" Feb 17 00:07:12 crc kubenswrapper[4690]: I0217 00:07:12.975584 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:12 crc kubenswrapper[4690]: I0217 00:07:12.975640 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:12 crc kubenswrapper[4690]: I0217 00:07:12.975683 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:12 crc kubenswrapper[4690]: I0217 00:07:12.975608 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:12 crc kubenswrapper[4690]: E0217 00:07:12.975767 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:12 crc kubenswrapper[4690]: E0217 00:07:12.975895 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:12 crc kubenswrapper[4690]: E0217 00:07:12.976055 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:12 crc kubenswrapper[4690]: E0217 00:07:12.976123 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:14 crc kubenswrapper[4690]: I0217 00:07:14.975829 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:14 crc kubenswrapper[4690]: I0217 00:07:14.975885 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:14 crc kubenswrapper[4690]: E0217 00:07:14.976003 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:14 crc kubenswrapper[4690]: I0217 00:07:14.976084 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:14 crc kubenswrapper[4690]: E0217 00:07:14.976185 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:14 crc kubenswrapper[4690]: I0217 00:07:14.976240 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:14 crc kubenswrapper[4690]: E0217 00:07:14.976335 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:14 crc kubenswrapper[4690]: E0217 00:07:14.976466 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:16 crc kubenswrapper[4690]: I0217 00:07:16.980806 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:16 crc kubenswrapper[4690]: I0217 00:07:16.980879 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:16 crc kubenswrapper[4690]: I0217 00:07:16.980833 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:16 crc kubenswrapper[4690]: I0217 00:07:16.980809 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:16 crc kubenswrapper[4690]: E0217 00:07:16.981466 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:16 crc kubenswrapper[4690]: E0217 00:07:16.981607 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:16 crc kubenswrapper[4690]: E0217 00:07:16.981784 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:16 crc kubenswrapper[4690]: E0217 00:07:16.983032 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:18 crc kubenswrapper[4690]: I0217 00:07:18.975674 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:18 crc kubenswrapper[4690]: I0217 00:07:18.975731 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:18 crc kubenswrapper[4690]: I0217 00:07:18.975761 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:18 crc kubenswrapper[4690]: I0217 00:07:18.975687 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:18 crc kubenswrapper[4690]: E0217 00:07:18.975891 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:18 crc kubenswrapper[4690]: E0217 00:07:18.976008 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:18 crc kubenswrapper[4690]: E0217 00:07:18.976153 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:18 crc kubenswrapper[4690]: E0217 00:07:18.976830 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:20 crc kubenswrapper[4690]: I0217 00:07:20.975999 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:20 crc kubenswrapper[4690]: I0217 00:07:20.976032 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:20 crc kubenswrapper[4690]: E0217 00:07:20.976504 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:20 crc kubenswrapper[4690]: I0217 00:07:20.976155 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:20 crc kubenswrapper[4690]: I0217 00:07:20.976079 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:20 crc kubenswrapper[4690]: E0217 00:07:20.976650 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:20 crc kubenswrapper[4690]: E0217 00:07:20.976807 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:20 crc kubenswrapper[4690]: E0217 00:07:20.976986 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.672243 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/1.log" Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.673039 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/0.log" Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.673087 4690 generic.go:334] "Generic (PLEG): container finished" podID="797b14df-f7d0-419b-95f8-f02f25409e66" containerID="543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae" exitCode=1 Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.673127 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vb4v5" event={"ID":"797b14df-f7d0-419b-95f8-f02f25409e66","Type":"ContainerDied","Data":"543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae"} Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.673169 4690 scope.go:117] "RemoveContainer" containerID="727c74af7c53fd7a093f0e9243e23461edf9babf1e980d2d3860f2ed24de66ff" Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.673752 4690 scope.go:117] "RemoveContainer" containerID="543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae" Feb 17 00:07:22 crc kubenswrapper[4690]: E0217 00:07:22.673998 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-vb4v5_openshift-multus(797b14df-f7d0-419b-95f8-f02f25409e66)\"" pod="openshift-multus/multus-vb4v5" podUID="797b14df-f7d0-419b-95f8-f02f25409e66" Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.975446 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.975511 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.975594 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:22 crc kubenswrapper[4690]: E0217 00:07:22.975637 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.975446 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:22 crc kubenswrapper[4690]: E0217 00:07:22.975865 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:22 crc kubenswrapper[4690]: E0217 00:07:22.975968 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:22 crc kubenswrapper[4690]: E0217 00:07:22.976510 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:22 crc kubenswrapper[4690]: I0217 00:07:22.977063 4690 scope.go:117] "RemoveContainer" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:07:23 crc kubenswrapper[4690]: I0217 00:07:23.678513 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/1.log" Feb 17 00:07:23 crc kubenswrapper[4690]: I0217 00:07:23.681639 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/3.log" Feb 17 00:07:23 crc kubenswrapper[4690]: I0217 00:07:23.686108 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerStarted","Data":"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65"} Feb 17 00:07:23 crc kubenswrapper[4690]: I0217 00:07:23.687255 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:07:23 crc kubenswrapper[4690]: I0217 00:07:23.732545 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podStartSLOduration=96.732525972 podStartE2EDuration="1m36.732525972s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:23.731145862 +0000 UTC m=+117.466858613" watchObservedRunningTime="2026-02-17 00:07:23.732525972 +0000 UTC m=+117.468238733" Feb 17 00:07:23 crc kubenswrapper[4690]: I0217 00:07:23.961926 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lz646"] Feb 17 00:07:23 crc kubenswrapper[4690]: I0217 00:07:23.962092 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:23 crc kubenswrapper[4690]: E0217 00:07:23.962226 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:24 crc kubenswrapper[4690]: I0217 00:07:24.976063 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:24 crc kubenswrapper[4690]: I0217 00:07:24.976167 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:24 crc kubenswrapper[4690]: E0217 00:07:24.976555 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:24 crc kubenswrapper[4690]: I0217 00:07:24.976256 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:24 crc kubenswrapper[4690]: E0217 00:07:24.976819 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:24 crc kubenswrapper[4690]: E0217 00:07:24.976901 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:25 crc kubenswrapper[4690]: I0217 00:07:25.975802 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:25 crc kubenswrapper[4690]: E0217 00:07:25.975999 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:26 crc kubenswrapper[4690]: I0217 00:07:26.975955 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:26 crc kubenswrapper[4690]: I0217 00:07:26.976192 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:26 crc kubenswrapper[4690]: I0217 00:07:26.976295 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:26 crc kubenswrapper[4690]: E0217 00:07:26.977193 4690 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 17 00:07:26 crc kubenswrapper[4690]: E0217 00:07:26.977960 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:26 crc kubenswrapper[4690]: E0217 00:07:26.978139 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:26 crc kubenswrapper[4690]: E0217 00:07:26.978258 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:27 crc kubenswrapper[4690]: E0217 00:07:27.081906 4690 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 17 00:07:27 crc kubenswrapper[4690]: I0217 00:07:27.994901 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:27 crc kubenswrapper[4690]: E0217 00:07:27.995557 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:28 crc kubenswrapper[4690]: I0217 00:07:28.975866 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:28 crc kubenswrapper[4690]: I0217 00:07:28.975908 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:28 crc kubenswrapper[4690]: E0217 00:07:28.976078 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:28 crc kubenswrapper[4690]: I0217 00:07:28.976124 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:28 crc kubenswrapper[4690]: E0217 00:07:28.976318 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:28 crc kubenswrapper[4690]: E0217 00:07:28.976436 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:29 crc kubenswrapper[4690]: I0217 00:07:29.975506 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:29 crc kubenswrapper[4690]: E0217 00:07:29.975693 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:30 crc kubenswrapper[4690]: I0217 00:07:30.975771 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:30 crc kubenswrapper[4690]: I0217 00:07:30.975793 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:30 crc kubenswrapper[4690]: E0217 00:07:30.976018 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:30 crc kubenswrapper[4690]: I0217 00:07:30.976115 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:30 crc kubenswrapper[4690]: E0217 00:07:30.976191 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:30 crc kubenswrapper[4690]: E0217 00:07:30.976306 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:31 crc kubenswrapper[4690]: I0217 00:07:31.975551 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:31 crc kubenswrapper[4690]: E0217 00:07:31.975763 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:32 crc kubenswrapper[4690]: E0217 00:07:32.083954 4690 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 17 00:07:32 crc kubenswrapper[4690]: I0217 00:07:32.975973 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:32 crc kubenswrapper[4690]: I0217 00:07:32.976023 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:32 crc kubenswrapper[4690]: E0217 00:07:32.976208 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:32 crc kubenswrapper[4690]: E0217 00:07:32.976311 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:32 crc kubenswrapper[4690]: I0217 00:07:32.977117 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:32 crc kubenswrapper[4690]: E0217 00:07:32.977463 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:33 crc kubenswrapper[4690]: I0217 00:07:33.975846 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:33 crc kubenswrapper[4690]: I0217 00:07:33.976267 4690 scope.go:117] "RemoveContainer" containerID="543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae" Feb 17 00:07:33 crc kubenswrapper[4690]: E0217 00:07:33.976403 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:34 crc kubenswrapper[4690]: I0217 00:07:34.727700 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/1.log" Feb 17 00:07:34 crc kubenswrapper[4690]: I0217 00:07:34.727791 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vb4v5" event={"ID":"797b14df-f7d0-419b-95f8-f02f25409e66","Type":"ContainerStarted","Data":"b9405af1a1a3ca87c5a1110233884fa8b8dad590f816aa1ac3af9d2e5dd32ff6"} Feb 17 00:07:34 crc kubenswrapper[4690]: I0217 00:07:34.975646 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:34 crc kubenswrapper[4690]: I0217 00:07:34.975647 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:34 crc kubenswrapper[4690]: E0217 00:07:34.975776 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:34 crc kubenswrapper[4690]: I0217 00:07:34.975647 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:34 crc kubenswrapper[4690]: E0217 00:07:34.975875 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:34 crc kubenswrapper[4690]: E0217 00:07:34.975939 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:35 crc kubenswrapper[4690]: I0217 00:07:35.975287 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:35 crc kubenswrapper[4690]: E0217 00:07:35.975526 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lz646" podUID="9050c3d5-4d74-4b57-afba-1dd177ce7983" Feb 17 00:07:36 crc kubenswrapper[4690]: I0217 00:07:36.975139 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:36 crc kubenswrapper[4690]: I0217 00:07:36.975289 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:36 crc kubenswrapper[4690]: I0217 00:07:36.976596 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:36 crc kubenswrapper[4690]: E0217 00:07:36.976587 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 00:07:36 crc kubenswrapper[4690]: E0217 00:07:36.976946 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 00:07:36 crc kubenswrapper[4690]: E0217 00:07:36.976974 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 00:07:37 crc kubenswrapper[4690]: I0217 00:07:37.975486 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:07:37 crc kubenswrapper[4690]: I0217 00:07:37.978160 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 17 00:07:37 crc kubenswrapper[4690]: I0217 00:07:37.979530 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 17 00:07:38 crc kubenswrapper[4690]: I0217 00:07:38.975689 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:38 crc kubenswrapper[4690]: I0217 00:07:38.975719 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:38 crc kubenswrapper[4690]: I0217 00:07:38.975891 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:38 crc kubenswrapper[4690]: I0217 00:07:38.978025 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 17 00:07:38 crc kubenswrapper[4690]: I0217 00:07:38.978043 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 17 00:07:38 crc kubenswrapper[4690]: I0217 00:07:38.978190 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 17 00:07:38 crc kubenswrapper[4690]: I0217 00:07:38.979735 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 17 00:07:42 crc kubenswrapper[4690]: I0217 00:07:42.034785 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.354963 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.409339 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vpd8p"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.409897 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.411600 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gzx5v"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.412452 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.413779 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vgzbg"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.414622 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.414632 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.414732 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.415931 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.418597 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.419186 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.420482 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.421812 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.422185 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.422838 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.423041 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.423423 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93685094-07ae-434e-a48c-258dc74730be-serving-cert\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.423848 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg8s5\" (UniqueName: \"kubernetes.io/projected/93685094-07ae-434e-a48c-258dc74730be-kube-api-access-rg8s5\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.423905 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93685094-07ae-434e-a48c-258dc74730be-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.423976 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93685094-07ae-434e-a48c-258dc74730be-service-ca-bundle\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.424012 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93685094-07ae-434e-a48c-258dc74730be-config\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.427907 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.432080 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.432286 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.432574 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.433113 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.433396 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.434382 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.434964 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.436885 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.439038 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.439671 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.441043 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.441839 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.441856 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.442193 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.442224 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.442255 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.443266 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.444064 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.444962 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.445395 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29521440-6zdhx"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.445634 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.445758 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.445922 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.445964 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29521440-6zdhx" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.449741 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-c6pc7"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.449879 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.449948 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.449984 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.450454 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.450604 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.452846 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.453294 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.453323 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.453989 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.455827 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.457001 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.464483 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.464640 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.464758 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.465128 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.465247 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.465283 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.465287 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.465250 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.465662 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.465790 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.468819 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.469089 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.478859 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.479091 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2j49q"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.479845 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vwd2q"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.480296 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-xflvn"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.480721 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xflvn" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.481128 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.481292 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.481478 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.481641 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.481958 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.482118 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.492523 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.492694 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.496208 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.496759 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.498458 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.501986 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.502383 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-glck9"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.502634 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.502927 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.503299 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.503543 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.506343 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.506698 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.506959 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.507157 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.507930 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.508175 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.508181 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.508311 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.515525 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.515920 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.516194 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.516584 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.516706 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.516929 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.517387 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h9fqv"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.517910 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.518571 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.537593 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.540456 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.541221 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93685094-07ae-434e-a48c-258dc74730be-config\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.541414 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93685094-07ae-434e-a48c-258dc74730be-serving-cert\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.541528 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.541617 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg8s5\" (UniqueName: \"kubernetes.io/projected/93685094-07ae-434e-a48c-258dc74730be-kube-api-access-rg8s5\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.541748 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt7lg\" (UniqueName: \"kubernetes.io/projected/8daa0570-a319-44d6-be28-c5717cb07575-kube-api-access-xt7lg\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.541878 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0936c3dd-a635-4721-8499-cd99755c8de8-images\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.542001 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-client-ca\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.542108 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0936c3dd-a635-4721-8499-cd99755c8de8-config\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.542226 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8daa0570-a319-44d6-be28-c5717cb07575-serving-cert\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.542342 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.542497 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-config\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.542606 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93685094-07ae-434e-a48c-258dc74730be-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.542718 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58xml\" (UniqueName: \"kubernetes.io/projected/0936c3dd-a635-4721-8499-cd99755c8de8-kube-api-access-58xml\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.542824 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0936c3dd-a635-4721-8499-cd99755c8de8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.542948 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93685094-07ae-434e-a48c-258dc74730be-service-ca-bundle\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.543481 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93685094-07ae-434e-a48c-258dc74730be-config\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.543479 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.543884 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93685094-07ae-434e-a48c-258dc74730be-service-ca-bundle\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.545020 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.542621 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.545238 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.545865 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93685094-07ae-434e-a48c-258dc74730be-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.556264 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557453 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557477 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557204 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.556335 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.556692 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.556938 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557718 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557010 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557201 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557259 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557293 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557306 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557353 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557386 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557415 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.557643 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.558071 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.558407 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.558548 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.559083 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93685094-07ae-434e-a48c-258dc74730be-serving-cert\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.559335 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.559667 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.559768 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.560658 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.560846 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.561033 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.561200 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.561040 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.561414 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.561455 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.561722 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.561943 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.563205 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.563426 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.563624 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.564044 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.569261 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.569637 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.569832 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.571231 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.575443 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.575861 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.575954 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.581913 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.582644 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-687jq"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.582775 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.583086 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.583500 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-g8jrf"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.584146 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.590488 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.591864 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.593641 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.593997 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pbbhc"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.595188 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.596296 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.595564 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.596587 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29521440-6zdhx"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.597712 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rdw2m"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.599020 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.599219 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.600127 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.600289 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.600932 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.601546 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.603242 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9pwzh"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.603796 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.604319 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.604981 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.606283 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gxcpk"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.607077 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.608293 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.609033 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gzx5v"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.609864 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.611113 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vpd8p"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.612581 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.613266 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.613904 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.614914 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2jr5f"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.615729 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.616204 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vgzbg"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.617291 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-4h7ff"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.618328 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4h7ff" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.618665 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.620317 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vwd2q"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.623316 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h9fqv"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.624505 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.625621 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.632659 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.633703 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.633888 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.639860 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.643927 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.643958 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.643983 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e069c9f2-7d28-4499-b432-c1bd329fcd3a-srv-cert\") pod \"catalog-operator-68c6474976-d67hq\" (UID: \"e069c9f2-7d28-4499-b432-c1bd329fcd3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644008 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8601b0b5-9982-4e47-98b5-f36b9ceb89ff-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9jxf8\" (UID: \"8601b0b5-9982-4e47-98b5-f36b9ceb89ff\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644030 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czjs9\" (UniqueName: \"kubernetes.io/projected/0903ddf6-73c1-4f80-99de-54b204d9df38-kube-api-access-czjs9\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644051 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e265f262-30e7-430d-b0f7-05db8db0c03d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2j49q\" (UID: \"e265f262-30e7-430d-b0f7-05db8db0c03d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644071 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644096 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrzml\" (UniqueName: \"kubernetes.io/projected/51b6988f-45a7-4faa-9f25-0c03e4fe316b-kube-api-access-hrzml\") pod \"control-plane-machine-set-operator-78cbb6b69f-7rclh\" (UID: \"51b6988f-45a7-4faa-9f25-0c03e4fe316b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644388 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acc970de-869b-4a13-8d0e-94370922b22f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6pwz\" (UID: \"acc970de-869b-4a13-8d0e-94370922b22f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644419 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/51b6988f-45a7-4faa-9f25-0c03e4fe316b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7rclh\" (UID: \"51b6988f-45a7-4faa-9f25-0c03e4fe316b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644443 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0903ddf6-73c1-4f80-99de-54b204d9df38-auth-proxy-config\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644465 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-policies\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644487 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpx5n\" (UniqueName: \"kubernetes.io/projected/b85e9f0e-5e1f-4d99-967e-1a3d6b12202c-kube-api-access-vpx5n\") pod \"service-ca-9c57cc56f-h9fqv\" (UID: \"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c\") " pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644541 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644563 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd0a26f5-e912-40a5-a41f-3754297b6f68-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644624 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hnpk\" (UniqueName: \"kubernetes.io/projected/e265f262-30e7-430d-b0f7-05db8db0c03d-kube-api-access-7hnpk\") pod \"openshift-config-operator-7777fb866f-2j49q\" (UID: \"e265f262-30e7-430d-b0f7-05db8db0c03d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644644 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644683 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644709 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b85e9f0e-5e1f-4d99-967e-1a3d6b12202c-signing-cabundle\") pod \"service-ca-9c57cc56f-h9fqv\" (UID: \"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c\") " pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644727 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg4bv\" (UniqueName: \"kubernetes.io/projected/0ded4d5d-1f0b-4caf-a498-73f7963446fb-kube-api-access-sg4bv\") pod \"dns-operator-744455d44c-vwd2q\" (UID: \"0ded4d5d-1f0b-4caf-a498-73f7963446fb\") " pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644773 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt7lg\" (UniqueName: \"kubernetes.io/projected/8daa0570-a319-44d6-be28-c5717cb07575-kube-api-access-xt7lg\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644796 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-metrics-certs\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644814 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cd0a26f5-e912-40a5-a41f-3754297b6f68-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644858 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cd0a26f5-e912-40a5-a41f-3754297b6f68-etcd-client\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644878 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97ljv\" (UniqueName: \"kubernetes.io/projected/ab15103a-6de2-47f7-8640-6c3253d2ee68-kube-api-access-97ljv\") pod \"service-ca-operator-777779d784-nxjhm\" (UID: \"ab15103a-6de2-47f7-8640-6c3253d2ee68\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644898 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btll4\" (UniqueName: \"kubernetes.io/projected/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-kube-api-access-btll4\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644942 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a94ddfaa-d52b-4ded-ad8b-c0732e08a983-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nstcz\" (UID: \"a94ddfaa-d52b-4ded-ad8b-c0732e08a983\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.644963 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wxst\" (UniqueName: \"kubernetes.io/projected/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-kube-api-access-9wxst\") pod \"image-pruner-29521440-6zdhx\" (UID: \"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6\") " pod="openshift-image-registry/image-pruner-29521440-6zdhx" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.645003 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cd0a26f5-e912-40a5-a41f-3754297b6f68-audit-policies\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.645035 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8bpg\" (UniqueName: \"kubernetes.io/projected/cd0a26f5-e912-40a5-a41f-3754297b6f68-kube-api-access-g8bpg\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.645053 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b85e9f0e-5e1f-4d99-967e-1a3d6b12202c-signing-key\") pod \"service-ca-9c57cc56f-h9fqv\" (UID: \"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c\") " pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.645095 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-client-ca\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.645115 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/21ad60ce-9b8d-4dfc-bc60-06901e688248-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.645138 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-default-certificate\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.645184 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a94ddfaa-d52b-4ded-ad8b-c0732e08a983-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nstcz\" (UID: \"a94ddfaa-d52b-4ded-ad8b-c0732e08a983\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646571 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/edcf1adb-8ea8-4b8f-86c6-45d882e984c1-srv-cert\") pod \"olm-operator-6b444d44fb-wkxlb\" (UID: \"edcf1adb-8ea8-4b8f-86c6-45d882e984c1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646614 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0936c3dd-a635-4721-8499-cd99755c8de8-images\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646659 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159fd3d3-b560-4f05-b24e-25611073d6b8-serving-cert\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646690 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/21ad60ce-9b8d-4dfc-bc60-06901e688248-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646712 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cd0a26f5-e912-40a5-a41f-3754297b6f68-encryption-config\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646736 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-config\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646759 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8601b0b5-9982-4e47-98b5-f36b9ceb89ff-proxy-tls\") pod \"machine-config-controller-84d6567774-9jxf8\" (UID: \"8601b0b5-9982-4e47-98b5-f36b9ceb89ff\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646782 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-stats-auth\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646808 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-client-ca\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646829 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ded4d5d-1f0b-4caf-a498-73f7963446fb-metrics-tls\") pod \"dns-operator-744455d44c-vwd2q\" (UID: \"0ded4d5d-1f0b-4caf-a498-73f7963446fb\") " pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646851 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbpr2\" (UniqueName: \"kubernetes.io/projected/a94ddfaa-d52b-4ded-ad8b-c0732e08a983-kube-api-access-cbpr2\") pod \"kube-storage-version-migrator-operator-b67b599dd-nstcz\" (UID: \"a94ddfaa-d52b-4ded-ad8b-c0732e08a983\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646874 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0903ddf6-73c1-4f80-99de-54b204d9df38-machine-approver-tls\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646897 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f6jp\" (UniqueName: \"kubernetes.io/projected/b7b8197f-4d82-4a0b-bed6-485dfaab9024-kube-api-access-2f6jp\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646918 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acc970de-869b-4a13-8d0e-94370922b22f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6pwz\" (UID: \"acc970de-869b-4a13-8d0e-94370922b22f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646945 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0936c3dd-a635-4721-8499-cd99755c8de8-config\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.646970 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h7f6\" (UniqueName: \"kubernetes.io/projected/159fd3d3-b560-4f05-b24e-25611073d6b8-kube-api-access-6h7f6\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647023 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647048 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9ksg\" (UniqueName: \"kubernetes.io/projected/50b867d9-7f2c-44b0-a432-95ae7e82f673-kube-api-access-s9ksg\") pod \"migrator-59844c95c7-pld4l\" (UID: \"50b867d9-7f2c-44b0-a432-95ae7e82f673\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647073 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7vt7\" (UniqueName: \"kubernetes.io/projected/edcf1adb-8ea8-4b8f-86c6-45d882e984c1-kube-api-access-t7vt7\") pod \"olm-operator-6b444d44fb-wkxlb\" (UID: \"edcf1adb-8ea8-4b8f-86c6-45d882e984c1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647100 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8daa0570-a319-44d6-be28-c5717cb07575-serving-cert\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647122 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6j9d\" (UniqueName: \"kubernetes.io/projected/8601b0b5-9982-4e47-98b5-f36b9ceb89ff-kube-api-access-t6j9d\") pod \"machine-config-controller-84d6567774-9jxf8\" (UID: \"8601b0b5-9982-4e47-98b5-f36b9ceb89ff\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647144 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e265f262-30e7-430d-b0f7-05db8db0c03d-serving-cert\") pod \"openshift-config-operator-7777fb866f-2j49q\" (UID: \"e265f262-30e7-430d-b0f7-05db8db0c03d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647176 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647279 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e069c9f2-7d28-4499-b432-c1bd329fcd3a-profile-collector-cert\") pod \"catalog-operator-68c6474976-d67hq\" (UID: \"e069c9f2-7d28-4499-b432-c1bd329fcd3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647308 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647332 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd0a26f5-e912-40a5-a41f-3754297b6f68-serving-cert\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647369 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab15103a-6de2-47f7-8640-6c3253d2ee68-config\") pod \"service-ca-operator-777779d784-nxjhm\" (UID: \"ab15103a-6de2-47f7-8640-6c3253d2ee68\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647392 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/edcf1adb-8ea8-4b8f-86c6-45d882e984c1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wkxlb\" (UID: \"edcf1adb-8ea8-4b8f-86c6-45d882e984c1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.647417 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0903ddf6-73c1-4f80-99de-54b204d9df38-config\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.648136 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-client-ca\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.648115 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.648188 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zskwr\" (UniqueName: \"kubernetes.io/projected/e069c9f2-7d28-4499-b432-c1bd329fcd3a-kube-api-access-zskwr\") pod \"catalog-operator-68c6474976-d67hq\" (UID: \"e069c9f2-7d28-4499-b432-c1bd329fcd3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.648325 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cd0a26f5-e912-40a5-a41f-3754297b6f68-audit-dir\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.648392 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab15103a-6de2-47f7-8640-6c3253d2ee68-serving-cert\") pod \"service-ca-operator-777779d784-nxjhm\" (UID: \"ab15103a-6de2-47f7-8640-6c3253d2ee68\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.648418 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-config\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.648881 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0936c3dd-a635-4721-8499-cd99755c8de8-config\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.648970 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58xml\" (UniqueName: \"kubernetes.io/projected/0936c3dd-a635-4721-8499-cd99755c8de8-kube-api-access-58xml\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.649034 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnmx9\" (UniqueName: \"kubernetes.io/projected/21ad60ce-9b8d-4dfc-bc60-06901e688248-kube-api-access-gnmx9\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.649077 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acc970de-869b-4a13-8d0e-94370922b22f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6pwz\" (UID: \"acc970de-869b-4a13-8d0e-94370922b22f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.649171 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0936c3dd-a635-4721-8499-cd99755c8de8-images\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.649228 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0936c3dd-a635-4721-8499-cd99755c8de8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.649338 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-service-ca-bundle\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.649524 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.649764 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21ad60ce-9b8d-4dfc-bc60-06901e688248-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.649898 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpgj4\" (UniqueName: \"kubernetes.io/projected/929321b2-58ac-4633-af87-da506a21c23a-kube-api-access-hpgj4\") pod \"downloads-7954f5f757-xflvn\" (UID: \"929321b2-58ac-4633-af87-da506a21c23a\") " pod="openshift-console/downloads-7954f5f757-xflvn" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.650035 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.650109 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-dir\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.650253 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-serviceca\") pod \"image-pruner-29521440-6zdhx\" (UID: \"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6\") " pod="openshift-image-registry/image-pruner-29521440-6zdhx" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.650329 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.651183 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-c6pc7"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.652074 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.653795 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.654377 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0936c3dd-a635-4721-8499-cd99755c8de8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.655726 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.657023 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-j28sv"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.658044 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.660074 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xflvn"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.660450 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-config\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.662944 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8daa0570-a319-44d6-be28-c5717cb07575-serving-cert\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.663034 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.663064 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pbbhc"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.666152 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2j49q"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.666287 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.666351 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-687jq"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.668512 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2jr5f"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.670114 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-g8jrf"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.671833 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.672460 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rdw2m"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.673325 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.675082 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.675929 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.677377 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.678655 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gxcpk"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.680727 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.682151 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.684057 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9pwzh"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.685663 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.686852 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.688284 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.689600 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.691291 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4h7ff"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.691422 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.693084 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-j28sv"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.694286 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-wgtxp"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.695198 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.696236 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bkbbg"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.697028 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.698447 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wgtxp"] Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.711878 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751258 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-config\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751291 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cd0a26f5-e912-40a5-a41f-3754297b6f68-encryption-config\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751310 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8601b0b5-9982-4e47-98b5-f36b9ceb89ff-proxy-tls\") pod \"machine-config-controller-84d6567774-9jxf8\" (UID: \"8601b0b5-9982-4e47-98b5-f36b9ceb89ff\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751326 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-stats-auth\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751344 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ded4d5d-1f0b-4caf-a498-73f7963446fb-metrics-tls\") pod \"dns-operator-744455d44c-vwd2q\" (UID: \"0ded4d5d-1f0b-4caf-a498-73f7963446fb\") " pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751376 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbpr2\" (UniqueName: \"kubernetes.io/projected/a94ddfaa-d52b-4ded-ad8b-c0732e08a983-kube-api-access-cbpr2\") pod \"kube-storage-version-migrator-operator-b67b599dd-nstcz\" (UID: \"a94ddfaa-d52b-4ded-ad8b-c0732e08a983\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751397 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0903ddf6-73c1-4f80-99de-54b204d9df38-machine-approver-tls\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751414 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f6jp\" (UniqueName: \"kubernetes.io/projected/b7b8197f-4d82-4a0b-bed6-485dfaab9024-kube-api-access-2f6jp\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751431 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acc970de-869b-4a13-8d0e-94370922b22f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6pwz\" (UID: \"acc970de-869b-4a13-8d0e-94370922b22f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751448 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h7f6\" (UniqueName: \"kubernetes.io/projected/159fd3d3-b560-4f05-b24e-25611073d6b8-kube-api-access-6h7f6\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751464 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751479 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9ksg\" (UniqueName: \"kubernetes.io/projected/50b867d9-7f2c-44b0-a432-95ae7e82f673-kube-api-access-s9ksg\") pod \"migrator-59844c95c7-pld4l\" (UID: \"50b867d9-7f2c-44b0-a432-95ae7e82f673\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751499 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7vt7\" (UniqueName: \"kubernetes.io/projected/edcf1adb-8ea8-4b8f-86c6-45d882e984c1-kube-api-access-t7vt7\") pod \"olm-operator-6b444d44fb-wkxlb\" (UID: \"edcf1adb-8ea8-4b8f-86c6-45d882e984c1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751517 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751541 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6j9d\" (UniqueName: \"kubernetes.io/projected/8601b0b5-9982-4e47-98b5-f36b9ceb89ff-kube-api-access-t6j9d\") pod \"machine-config-controller-84d6567774-9jxf8\" (UID: \"8601b0b5-9982-4e47-98b5-f36b9ceb89ff\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751558 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e265f262-30e7-430d-b0f7-05db8db0c03d-serving-cert\") pod \"openshift-config-operator-7777fb866f-2j49q\" (UID: \"e265f262-30e7-430d-b0f7-05db8db0c03d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751575 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e069c9f2-7d28-4499-b432-c1bd329fcd3a-profile-collector-cert\") pod \"catalog-operator-68c6474976-d67hq\" (UID: \"e069c9f2-7d28-4499-b432-c1bd329fcd3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751596 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d6d4f41-49a8-4ddd-acfb-1695db85409c-config-volume\") pod \"collect-profiles-29521440-8z9bb\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751611 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/edcf1adb-8ea8-4b8f-86c6-45d882e984c1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wkxlb\" (UID: \"edcf1adb-8ea8-4b8f-86c6-45d882e984c1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751628 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd0a26f5-e912-40a5-a41f-3754297b6f68-serving-cert\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751643 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab15103a-6de2-47f7-8640-6c3253d2ee68-config\") pod \"service-ca-operator-777779d784-nxjhm\" (UID: \"ab15103a-6de2-47f7-8640-6c3253d2ee68\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751657 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d6d4f41-49a8-4ddd-acfb-1695db85409c-secret-volume\") pod \"collect-profiles-29521440-8z9bb\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751675 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cd0a26f5-e912-40a5-a41f-3754297b6f68-audit-dir\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751694 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0903ddf6-73c1-4f80-99de-54b204d9df38-config\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751715 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751739 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zskwr\" (UniqueName: \"kubernetes.io/projected/e069c9f2-7d28-4499-b432-c1bd329fcd3a-kube-api-access-zskwr\") pod \"catalog-operator-68c6474976-d67hq\" (UID: \"e069c9f2-7d28-4499-b432-c1bd329fcd3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751760 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab15103a-6de2-47f7-8640-6c3253d2ee68-serving-cert\") pod \"service-ca-operator-777779d784-nxjhm\" (UID: \"ab15103a-6de2-47f7-8640-6c3253d2ee68\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751793 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnmx9\" (UniqueName: \"kubernetes.io/projected/21ad60ce-9b8d-4dfc-bc60-06901e688248-kube-api-access-gnmx9\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751816 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acc970de-869b-4a13-8d0e-94370922b22f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6pwz\" (UID: \"acc970de-869b-4a13-8d0e-94370922b22f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751846 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-service-ca-bundle\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751867 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751887 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21ad60ce-9b8d-4dfc-bc60-06901e688248-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751902 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpgj4\" (UniqueName: \"kubernetes.io/projected/929321b2-58ac-4633-af87-da506a21c23a-kube-api-access-hpgj4\") pod \"downloads-7954f5f757-xflvn\" (UID: \"929321b2-58ac-4633-af87-da506a21c23a\") " pod="openshift-console/downloads-7954f5f757-xflvn" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751920 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f06f4ec3-86a3-429b-8920-03dd8e7a4569-serving-cert\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751935 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751951 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-dir\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.751975 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-serviceca\") pod \"image-pruner-29521440-6zdhx\" (UID: \"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6\") " pod="openshift-image-registry/image-pruner-29521440-6zdhx" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752021 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8601b0b5-9982-4e47-98b5-f36b9ceb89ff-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9jxf8\" (UID: \"8601b0b5-9982-4e47-98b5-f36b9ceb89ff\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752045 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czjs9\" (UniqueName: \"kubernetes.io/projected/0903ddf6-73c1-4f80-99de-54b204d9df38-kube-api-access-czjs9\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752067 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752087 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752105 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e069c9f2-7d28-4499-b432-c1bd329fcd3a-srv-cert\") pod \"catalog-operator-68c6474976-d67hq\" (UID: \"e069c9f2-7d28-4499-b432-c1bd329fcd3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752128 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s6v8\" (UniqueName: \"kubernetes.io/projected/f06f4ec3-86a3-429b-8920-03dd8e7a4569-kube-api-access-4s6v8\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752166 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e265f262-30e7-430d-b0f7-05db8db0c03d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2j49q\" (UID: \"e265f262-30e7-430d-b0f7-05db8db0c03d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752192 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752223 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrzml\" (UniqueName: \"kubernetes.io/projected/51b6988f-45a7-4faa-9f25-0c03e4fe316b-kube-api-access-hrzml\") pod \"control-plane-machine-set-operator-78cbb6b69f-7rclh\" (UID: \"51b6988f-45a7-4faa-9f25-0c03e4fe316b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752249 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acc970de-869b-4a13-8d0e-94370922b22f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6pwz\" (UID: \"acc970de-869b-4a13-8d0e-94370922b22f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752273 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/51b6988f-45a7-4faa-9f25-0c03e4fe316b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7rclh\" (UID: \"51b6988f-45a7-4faa-9f25-0c03e4fe316b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752295 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0903ddf6-73c1-4f80-99de-54b204d9df38-auth-proxy-config\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752309 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-policies\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752325 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpx5n\" (UniqueName: \"kubernetes.io/projected/b85e9f0e-5e1f-4d99-967e-1a3d6b12202c-kube-api-access-vpx5n\") pod \"service-ca-9c57cc56f-h9fqv\" (UID: \"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c\") " pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752370 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752385 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd0a26f5-e912-40a5-a41f-3754297b6f68-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752402 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hnpk\" (UniqueName: \"kubernetes.io/projected/e265f262-30e7-430d-b0f7-05db8db0c03d-kube-api-access-7hnpk\") pod \"openshift-config-operator-7777fb866f-2j49q\" (UID: \"e265f262-30e7-430d-b0f7-05db8db0c03d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752417 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752432 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752449 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg4bv\" (UniqueName: \"kubernetes.io/projected/0ded4d5d-1f0b-4caf-a498-73f7963446fb-kube-api-access-sg4bv\") pod \"dns-operator-744455d44c-vwd2q\" (UID: \"0ded4d5d-1f0b-4caf-a498-73f7963446fb\") " pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752464 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b85e9f0e-5e1f-4d99-967e-1a3d6b12202c-signing-cabundle\") pod \"service-ca-9c57cc56f-h9fqv\" (UID: \"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c\") " pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752485 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-metrics-certs\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752499 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cd0a26f5-e912-40a5-a41f-3754297b6f68-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752514 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btll4\" (UniqueName: \"kubernetes.io/projected/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-kube-api-access-btll4\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752528 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cd0a26f5-e912-40a5-a41f-3754297b6f68-etcd-client\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752543 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97ljv\" (UniqueName: \"kubernetes.io/projected/ab15103a-6de2-47f7-8640-6c3253d2ee68-kube-api-access-97ljv\") pod \"service-ca-operator-777779d784-nxjhm\" (UID: \"ab15103a-6de2-47f7-8640-6c3253d2ee68\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752557 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cd0a26f5-e912-40a5-a41f-3754297b6f68-audit-policies\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752579 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a94ddfaa-d52b-4ded-ad8b-c0732e08a983-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nstcz\" (UID: \"a94ddfaa-d52b-4ded-ad8b-c0732e08a983\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752595 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wxst\" (UniqueName: \"kubernetes.io/projected/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-kube-api-access-9wxst\") pod \"image-pruner-29521440-6zdhx\" (UID: \"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6\") " pod="openshift-image-registry/image-pruner-29521440-6zdhx" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752610 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8bpg\" (UniqueName: \"kubernetes.io/projected/cd0a26f5-e912-40a5-a41f-3754297b6f68-kube-api-access-g8bpg\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752624 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b85e9f0e-5e1f-4d99-967e-1a3d6b12202c-signing-key\") pod \"service-ca-9c57cc56f-h9fqv\" (UID: \"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c\") " pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752641 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5q78\" (UniqueName: \"kubernetes.io/projected/0d6d4f41-49a8-4ddd-acfb-1695db85409c-kube-api-access-c5q78\") pod \"collect-profiles-29521440-8z9bb\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752657 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-client-ca\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752672 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/21ad60ce-9b8d-4dfc-bc60-06901e688248-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752686 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f06f4ec3-86a3-429b-8920-03dd8e7a4569-config\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752709 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-default-certificate\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752727 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a94ddfaa-d52b-4ded-ad8b-c0732e08a983-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nstcz\" (UID: \"a94ddfaa-d52b-4ded-ad8b-c0732e08a983\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752743 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/edcf1adb-8ea8-4b8f-86c6-45d882e984c1-srv-cert\") pod \"olm-operator-6b444d44fb-wkxlb\" (UID: \"edcf1adb-8ea8-4b8f-86c6-45d882e984c1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752759 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159fd3d3-b560-4f05-b24e-25611073d6b8-serving-cert\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752776 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/21ad60ce-9b8d-4dfc-bc60-06901e688248-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.752792 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f06f4ec3-86a3-429b-8920-03dd8e7a4569-trusted-ca\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.753063 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-config\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.754218 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cd0a26f5-e912-40a5-a41f-3754297b6f68-encryption-config\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.754229 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e265f262-30e7-430d-b0f7-05db8db0c03d-serving-cert\") pod \"openshift-config-operator-7777fb866f-2j49q\" (UID: \"e265f262-30e7-430d-b0f7-05db8db0c03d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.754646 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.754938 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0903ddf6-73c1-4f80-99de-54b204d9df38-machine-approver-tls\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.754956 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acc970de-869b-4a13-8d0e-94370922b22f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6pwz\" (UID: \"acc970de-869b-4a13-8d0e-94370922b22f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.755110 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ded4d5d-1f0b-4caf-a498-73f7963446fb-metrics-tls\") pod \"dns-operator-744455d44c-vwd2q\" (UID: \"0ded4d5d-1f0b-4caf-a498-73f7963446fb\") " pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.755211 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-dir\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.755678 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd0a26f5-e912-40a5-a41f-3754297b6f68-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.755898 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-stats-auth\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.756155 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8601b0b5-9982-4e47-98b5-f36b9ceb89ff-proxy-tls\") pod \"machine-config-controller-84d6567774-9jxf8\" (UID: \"8601b0b5-9982-4e47-98b5-f36b9ceb89ff\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.756876 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.757010 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-policies\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.757271 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.757475 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0903ddf6-73c1-4f80-99de-54b204d9df38-auth-proxy-config\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.757504 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acc970de-869b-4a13-8d0e-94370922b22f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6pwz\" (UID: \"acc970de-869b-4a13-8d0e-94370922b22f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.757710 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0903ddf6-73c1-4f80-99de-54b204d9df38-config\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.757988 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cd0a26f5-e912-40a5-a41f-3754297b6f68-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.758623 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21ad60ce-9b8d-4dfc-bc60-06901e688248-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.758847 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-serviceca\") pod \"image-pruner-29521440-6zdhx\" (UID: \"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6\") " pod="openshift-image-registry/image-pruner-29521440-6zdhx" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.759088 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg8s5\" (UniqueName: \"kubernetes.io/projected/93685094-07ae-434e-a48c-258dc74730be-kube-api-access-rg8s5\") pod \"authentication-operator-69f744f599-vpd8p\" (UID: \"93685094-07ae-434e-a48c-258dc74730be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.759179 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab15103a-6de2-47f7-8640-6c3253d2ee68-config\") pod \"service-ca-operator-777779d784-nxjhm\" (UID: \"ab15103a-6de2-47f7-8640-6c3253d2ee68\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.759773 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8601b0b5-9982-4e47-98b5-f36b9ceb89ff-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9jxf8\" (UID: \"8601b0b5-9982-4e47-98b5-f36b9ceb89ff\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.760139 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-service-ca-bundle\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.760491 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.760526 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-client-ca\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.761008 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cd0a26f5-e912-40a5-a41f-3754297b6f68-audit-policies\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.761447 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/edcf1adb-8ea8-4b8f-86c6-45d882e984c1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wkxlb\" (UID: \"edcf1adb-8ea8-4b8f-86c6-45d882e984c1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.761720 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b85e9f0e-5e1f-4d99-967e-1a3d6b12202c-signing-cabundle\") pod \"service-ca-9c57cc56f-h9fqv\" (UID: \"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c\") " pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.762253 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a94ddfaa-d52b-4ded-ad8b-c0732e08a983-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nstcz\" (UID: \"a94ddfaa-d52b-4ded-ad8b-c0732e08a983\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.763667 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e069c9f2-7d28-4499-b432-c1bd329fcd3a-profile-collector-cert\") pod \"catalog-operator-68c6474976-d67hq\" (UID: \"e069c9f2-7d28-4499-b432-c1bd329fcd3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.763744 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cd0a26f5-e912-40a5-a41f-3754297b6f68-audit-dir\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.763675 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.763752 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e069c9f2-7d28-4499-b432-c1bd329fcd3a-srv-cert\") pod \"catalog-operator-68c6474976-d67hq\" (UID: \"e069c9f2-7d28-4499-b432-c1bd329fcd3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.763996 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e265f262-30e7-430d-b0f7-05db8db0c03d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2j49q\" (UID: \"e265f262-30e7-430d-b0f7-05db8db0c03d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.764008 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.764689 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.764981 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.765339 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/51b6988f-45a7-4faa-9f25-0c03e4fe316b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7rclh\" (UID: \"51b6988f-45a7-4faa-9f25-0c03e4fe316b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.765843 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab15103a-6de2-47f7-8640-6c3253d2ee68-serving-cert\") pod \"service-ca-operator-777779d784-nxjhm\" (UID: \"ab15103a-6de2-47f7-8640-6c3253d2ee68\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.766087 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.766237 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.767234 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a94ddfaa-d52b-4ded-ad8b-c0732e08a983-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nstcz\" (UID: \"a94ddfaa-d52b-4ded-ad8b-c0732e08a983\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.767562 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-default-certificate\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.767749 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.767847 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/edcf1adb-8ea8-4b8f-86c6-45d882e984c1-srv-cert\") pod \"olm-operator-6b444d44fb-wkxlb\" (UID: \"edcf1adb-8ea8-4b8f-86c6-45d882e984c1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.768033 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd0a26f5-e912-40a5-a41f-3754297b6f68-serving-cert\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.768328 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b85e9f0e-5e1f-4d99-967e-1a3d6b12202c-signing-key\") pod \"service-ca-9c57cc56f-h9fqv\" (UID: \"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c\") " pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.768921 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/21ad60ce-9b8d-4dfc-bc60-06901e688248-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.769538 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159fd3d3-b560-4f05-b24e-25611073d6b8-serving-cert\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.770209 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-metrics-certs\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.772038 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.782862 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cd0a26f5-e912-40a5-a41f-3754297b6f68-etcd-client\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.795307 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.812065 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.831884 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.852426 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.853743 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f06f4ec3-86a3-429b-8920-03dd8e7a4569-serving-cert\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.853937 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s6v8\" (UniqueName: \"kubernetes.io/projected/f06f4ec3-86a3-429b-8920-03dd8e7a4569-kube-api-access-4s6v8\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.854429 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5q78\" (UniqueName: \"kubernetes.io/projected/0d6d4f41-49a8-4ddd-acfb-1695db85409c-kube-api-access-c5q78\") pod \"collect-profiles-29521440-8z9bb\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.854473 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f06f4ec3-86a3-429b-8920-03dd8e7a4569-config\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.854498 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f06f4ec3-86a3-429b-8920-03dd8e7a4569-trusted-ca\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.854572 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d6d4f41-49a8-4ddd-acfb-1695db85409c-config-volume\") pod \"collect-profiles-29521440-8z9bb\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.854600 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d6d4f41-49a8-4ddd-acfb-1695db85409c-secret-volume\") pod \"collect-profiles-29521440-8z9bb\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.855485 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d6d4f41-49a8-4ddd-acfb-1695db85409c-config-volume\") pod \"collect-profiles-29521440-8z9bb\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.858050 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d6d4f41-49a8-4ddd-acfb-1695db85409c-secret-volume\") pod \"collect-profiles-29521440-8z9bb\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.872154 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.893574 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.912657 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.932635 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.952462 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.971864 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 17 00:07:46 crc kubenswrapper[4690]: I0217 00:07:46.993069 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.012346 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.038809 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.048297 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.052202 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.074129 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.093501 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.112905 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.133265 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.152201 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.174990 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.193328 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.212689 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.231812 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.253549 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.272919 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.292974 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.313475 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.326335 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vpd8p"] Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.332577 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: W0217 00:07:47.341437 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93685094_07ae_434e_a48c_258dc74730be.slice/crio-b99495889af76e160e255a9d2c4dc60ace2fa0564092275f6d8ef7b40cafdb67 WatchSource:0}: Error finding container b99495889af76e160e255a9d2c4dc60ace2fa0564092275f6d8ef7b40cafdb67: Status 404 returned error can't find the container with id b99495889af76e160e255a9d2c4dc60ace2fa0564092275f6d8ef7b40cafdb67 Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.352478 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.372709 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.392889 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.423647 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.426544 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f06f4ec3-86a3-429b-8920-03dd8e7a4569-trusted-ca\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.432587 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.453165 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.456026 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f06f4ec3-86a3-429b-8920-03dd8e7a4569-config\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.474756 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.494286 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.513671 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.522800 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f06f4ec3-86a3-429b-8920-03dd8e7a4569-serving-cert\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.534075 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.553319 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.572169 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.593965 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.610675 4690 request.go:700] Waited for 1.009575803s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager-operator/secrets?fieldSelector=metadata.name%3Dkube-controller-manager-operator-dockercfg-gkqpw&limit=500&resourceVersion=0 Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.613185 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.633684 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.652836 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.673076 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.693663 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.712892 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.733624 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.753564 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.773144 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.779178 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" event={"ID":"93685094-07ae-434e-a48c-258dc74730be","Type":"ContainerStarted","Data":"b74916074e03ad7c32a469b1e3ed4f858e519ea78fa8ed57a8a5b50d8b8ad01e"} Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.779246 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" event={"ID":"93685094-07ae-434e-a48c-258dc74730be","Type":"ContainerStarted","Data":"b99495889af76e160e255a9d2c4dc60ace2fa0564092275f6d8ef7b40cafdb67"} Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.799391 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.813488 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.833493 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.852302 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.872860 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.892498 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.912976 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.932635 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.952764 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 17 00:07:47 crc kubenswrapper[4690]: I0217 00:07:47.972735 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.002339 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.012741 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.033050 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.052987 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.074290 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.092887 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.124835 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.133521 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.152864 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.172183 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.192927 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.212959 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.233298 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.252705 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.273401 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.292669 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.312572 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.367951 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt7lg\" (UniqueName: \"kubernetes.io/projected/8daa0570-a319-44d6-be28-c5717cb07575-kube-api-access-xt7lg\") pod \"controller-manager-879f6c89f-gzx5v\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.372476 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58xml\" (UniqueName: \"kubernetes.io/projected/0936c3dd-a635-4721-8499-cd99755c8de8-kube-api-access-58xml\") pod \"machine-api-operator-5694c8668f-vgzbg\" (UID: \"0936c3dd-a635-4721-8499-cd99755c8de8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.373008 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.393122 4690 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.411850 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.452159 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.473078 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.492687 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.513130 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.533032 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.552093 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.556988 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.607074 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h7f6\" (UniqueName: \"kubernetes.io/projected/159fd3d3-b560-4f05-b24e-25611073d6b8-kube-api-access-6h7f6\") pod \"route-controller-manager-6576b87f9c-zw657\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.613404 4690 request.go:700] Waited for 1.860059056s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.617329 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.629101 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbpr2\" (UniqueName: \"kubernetes.io/projected/a94ddfaa-d52b-4ded-ad8b-c0732e08a983-kube-api-access-cbpr2\") pod \"kube-storage-version-migrator-operator-b67b599dd-nstcz\" (UID: \"a94ddfaa-d52b-4ded-ad8b-c0732e08a983\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.658342 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.663154 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f6jp\" (UniqueName: \"kubernetes.io/projected/b7b8197f-4d82-4a0b-bed6-485dfaab9024-kube-api-access-2f6jp\") pod \"oauth-openshift-558db77b4-c6pc7\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.700780 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7vt7\" (UniqueName: \"kubernetes.io/projected/edcf1adb-8ea8-4b8f-86c6-45d882e984c1-kube-api-access-t7vt7\") pod \"olm-operator-6b444d44fb-wkxlb\" (UID: \"edcf1adb-8ea8-4b8f-86c6-45d882e984c1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.706635 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9ksg\" (UniqueName: \"kubernetes.io/projected/50b867d9-7f2c-44b0-a432-95ae7e82f673-kube-api-access-s9ksg\") pod \"migrator-59844c95c7-pld4l\" (UID: \"50b867d9-7f2c-44b0-a432-95ae7e82f673\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.718737 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.719710 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hnpk\" (UniqueName: \"kubernetes.io/projected/e265f262-30e7-430d-b0f7-05db8db0c03d-kube-api-access-7hnpk\") pod \"openshift-config-operator-7777fb866f-2j49q\" (UID: \"e265f262-30e7-430d-b0f7-05db8db0c03d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.729741 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpx5n\" (UniqueName: \"kubernetes.io/projected/b85e9f0e-5e1f-4d99-967e-1a3d6b12202c-kube-api-access-vpx5n\") pod \"service-ca-9c57cc56f-h9fqv\" (UID: \"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c\") " pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.739960 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.761819 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnmx9\" (UniqueName: \"kubernetes.io/projected/21ad60ce-9b8d-4dfc-bc60-06901e688248-kube-api-access-gnmx9\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.771331 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zskwr\" (UniqueName: \"kubernetes.io/projected/e069c9f2-7d28-4499-b432-c1bd329fcd3a-kube-api-access-zskwr\") pod \"catalog-operator-68c6474976-d67hq\" (UID: \"e069c9f2-7d28-4499-b432-c1bd329fcd3a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.774266 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.789849 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6j9d\" (UniqueName: \"kubernetes.io/projected/8601b0b5-9982-4e47-98b5-f36b9ceb89ff-kube-api-access-t6j9d\") pod \"machine-config-controller-84d6567774-9jxf8\" (UID: \"8601b0b5-9982-4e47-98b5-f36b9ceb89ff\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.810328 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpgj4\" (UniqueName: \"kubernetes.io/projected/929321b2-58ac-4633-af87-da506a21c23a-kube-api-access-hpgj4\") pod \"downloads-7954f5f757-xflvn\" (UID: \"929321b2-58ac-4633-af87-da506a21c23a\") " pod="openshift-console/downloads-7954f5f757-xflvn" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.810680 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.820660 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.829995 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gzx5v"] Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.834604 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acc970de-869b-4a13-8d0e-94370922b22f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6pwz\" (UID: \"acc970de-869b-4a13-8d0e-94370922b22f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.843110 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.848784 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.853181 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97ljv\" (UniqueName: \"kubernetes.io/projected/ab15103a-6de2-47f7-8640-6c3253d2ee68-kube-api-access-97ljv\") pod \"service-ca-operator-777779d784-nxjhm\" (UID: \"ab15103a-6de2-47f7-8640-6c3253d2ee68\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.885383 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czjs9\" (UniqueName: \"kubernetes.io/projected/0903ddf6-73c1-4f80-99de-54b204d9df38-kube-api-access-czjs9\") pod \"machine-approver-56656f9798-bdbdq\" (UID: \"0903ddf6-73c1-4f80-99de-54b204d9df38\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.892845 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg4bv\" (UniqueName: \"kubernetes.io/projected/0ded4d5d-1f0b-4caf-a498-73f7963446fb-kube-api-access-sg4bv\") pod \"dns-operator-744455d44c-vwd2q\" (UID: \"0ded4d5d-1f0b-4caf-a498-73f7963446fb\") " pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.904719 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.907277 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8bpg\" (UniqueName: \"kubernetes.io/projected/cd0a26f5-e912-40a5-a41f-3754297b6f68-kube-api-access-g8bpg\") pod \"apiserver-7bbb656c7d-8sp7g\" (UID: \"cd0a26f5-e912-40a5-a41f-3754297b6f68\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.909105 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vgzbg"] Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.930813 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wxst\" (UniqueName: \"kubernetes.io/projected/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-kube-api-access-9wxst\") pod \"image-pruner-29521440-6zdhx\" (UID: \"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6\") " pod="openshift-image-registry/image-pruner-29521440-6zdhx" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.946744 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrzml\" (UniqueName: \"kubernetes.io/projected/51b6988f-45a7-4faa-9f25-0c03e4fe316b-kube-api-access-hrzml\") pod \"control-plane-machine-set-operator-78cbb6b69f-7rclh\" (UID: \"51b6988f-45a7-4faa-9f25-0c03e4fe316b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.958145 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657"] Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.966422 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btll4\" (UniqueName: \"kubernetes.io/projected/f56490e5-1c42-4126-b7f5-cd7ef73b32bd-kube-api-access-btll4\") pod \"router-default-5444994796-glck9\" (UID: \"f56490e5-1c42-4126-b7f5-cd7ef73b32bd\") " pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.980707 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.992223 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:48 crc kubenswrapper[4690]: I0217 00:07:48.996531 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/21ad60ce-9b8d-4dfc-bc60-06901e688248-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8f7ps\" (UID: \"21ad60ce-9b8d-4dfc-bc60-06901e688248\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.000891 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29521440-6zdhx" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.010204 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s6v8\" (UniqueName: \"kubernetes.io/projected/f06f4ec3-86a3-429b-8920-03dd8e7a4569-kube-api-access-4s6v8\") pod \"console-operator-58897d9998-pbbhc\" (UID: \"f06f4ec3-86a3-429b-8920-03dd8e7a4569\") " pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.021911 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.026331 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xflvn" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.043754 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5q78\" (UniqueName: \"kubernetes.io/projected/0d6d4f41-49a8-4ddd-acfb-1695db85409c-kube-api-access-c5q78\") pod \"collect-profiles-29521440-8z9bb\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.052853 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.085112 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2j49q"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.085289 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.092656 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.099825 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.099858 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-serving-cert\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.099878 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w65tp\" (UniqueName: \"kubernetes.io/projected/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-kube-api-access-w65tp\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.099894 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73b144f3-32f0-47ca-96dc-54eab0196a50-trusted-ca\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.099923 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.099956 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5604a4e2-1b7b-40a6-9c33-38469fb706b4-etcd-ca\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.099970 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6070dad-196d-4179-858c-4f75ddb71ebc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-66x5j\" (UID: \"b6070dad-196d-4179-858c-4f75ddb71ebc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.099985 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhdtv\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-kube-api-access-jhdtv\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100000 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-oauth-serving-cert\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100018 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j992c\" (UID: \"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100059 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkvtw\" (UniqueName: \"kubernetes.io/projected/e4268b59-7563-4783-9957-2d1656b4ddec-kube-api-access-rkvtw\") pod \"ingress-canary-4h7ff\" (UID: \"e4268b59-7563-4783-9957-2d1656b4ddec\") " pod="openshift-ingress-canary/ingress-canary-4h7ff" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100097 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-csi-data-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100111 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/56c57ff0-baf4-483e-8763-a3ba52c0130e-tmpfs\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100125 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gxcpk\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100142 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7sng\" (UniqueName: \"kubernetes.io/projected/42491e41-75ad-4087-90ea-69ec8d173716-kube-api-access-c7sng\") pod \"multus-admission-controller-857f4d67dd-2jr5f\" (UID: \"42491e41-75ad-4087-90ea-69ec8d173716\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100159 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6070dad-196d-4179-858c-4f75ddb71ebc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-66x5j\" (UID: \"b6070dad-196d-4179-858c-4f75ddb71ebc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100194 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/56c57ff0-baf4-483e-8763-a3ba52c0130e-apiservice-cert\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100227 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/56c57ff0-baf4-483e-8763-a3ba52c0130e-webhook-cert\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100242 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/983bb1a5-259a-4d98-8c43-4d91d56825bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wslzr\" (UID: \"983bb1a5-259a-4d98-8c43-4d91d56825bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100259 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5wgb\" (UniqueName: \"kubernetes.io/projected/b6070dad-196d-4179-858c-4f75ddb71ebc-kube-api-access-m5wgb\") pod \"openshift-controller-manager-operator-756b6f6bc6-66x5j\" (UID: \"b6070dad-196d-4179-858c-4f75ddb71ebc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100273 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xwpp\" (UniqueName: \"kubernetes.io/projected/5604a4e2-1b7b-40a6-9c33-38469fb706b4-kube-api-access-8xwpp\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100297 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl6c4\" (UniqueName: \"kubernetes.io/projected/cfb0993f-dcb1-48a5-a415-9ce91bdf75ed-kube-api-access-jl6c4\") pod \"openshift-apiserver-operator-796bbdcf4f-zjwj9\" (UID: \"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100312 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-console-oauth-config\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100329 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j992c\" (UID: \"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100398 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-mountpoint-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.100998 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.101304 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:49.601289369 +0000 UTC m=+143.337002120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.108520 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-node-pullsecrets\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.108610 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5604a4e2-1b7b-40a6-9c33-38469fb706b4-etcd-service-ca\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.108653 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gxcpk\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.108676 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7vtf\" (UniqueName: \"kubernetes.io/projected/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-kube-api-access-d7vtf\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.108698 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73b144f3-32f0-47ca-96dc-54eab0196a50-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.108719 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p67k\" (UniqueName: \"kubernetes.io/projected/b0373e24-779c-47ec-b8d7-d4152b995948-kube-api-access-5p67k\") pod \"cluster-samples-operator-665b6dd947-p2b72\" (UID: \"b0373e24-779c-47ec-b8d7-d4152b995948\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.108779 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac4bf154-96bf-4052-b8e6-371ede515470-config\") pod \"kube-apiserver-operator-766d6c64bb-xkbpk\" (UID: \"ac4bf154-96bf-4052-b8e6-371ede515470\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.109091 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac4bf154-96bf-4052-b8e6-371ede515470-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xkbpk\" (UID: \"ac4bf154-96bf-4052-b8e6-371ede515470\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.109327 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-socket-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.109369 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-registration-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.109394 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-audit-dir\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.109742 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.109777 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-console-config\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.109858 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-config\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110149 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggvrj\" (UniqueName: \"kubernetes.io/projected/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-kube-api-access-ggvrj\") pod \"marketplace-operator-79b997595-gxcpk\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110195 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z24n2\" (UniqueName: \"kubernetes.io/projected/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-kube-api-access-z24n2\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110256 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktd5k\" (UniqueName: \"kubernetes.io/projected/73b144f3-32f0-47ca-96dc-54eab0196a50-kube-api-access-ktd5k\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110298 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e4268b59-7563-4783-9957-2d1656b4ddec-cert\") pod \"ingress-canary-4h7ff\" (UID: \"e4268b59-7563-4783-9957-2d1656b4ddec\") " pod="openshift-ingress-canary/ingress-canary-4h7ff" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110322 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5604a4e2-1b7b-40a6-9c33-38469fb706b4-etcd-client\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110337 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac4bf154-96bf-4052-b8e6-371ede515470-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xkbpk\" (UID: \"ac4bf154-96bf-4052-b8e6-371ede515470\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110398 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5604a4e2-1b7b-40a6-9c33-38469fb706b4-serving-cert\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110429 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg2f4\" (UniqueName: \"kubernetes.io/projected/56c57ff0-baf4-483e-8763-a3ba52c0130e-kube-api-access-fg2f4\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110454 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-trusted-ca\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110498 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b0373e24-779c-47ec-b8d7-d4152b995948-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p2b72\" (UID: \"b0373e24-779c-47ec-b8d7-d4152b995948\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110520 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-images\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110535 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-etcd-client\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110559 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-trusted-ca-bundle\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110575 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/42491e41-75ad-4087-90ea-69ec8d173716-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2jr5f\" (UID: \"42491e41-75ad-4087-90ea-69ec8d173716\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110589 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/73b144f3-32f0-47ca-96dc-54eab0196a50-metrics-tls\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110622 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-tls\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110636 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-etcd-serving-ca\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110652 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5604a4e2-1b7b-40a6-9c33-38469fb706b4-config\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110690 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-certificates\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110707 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110723 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6q2h\" (UniqueName: \"kubernetes.io/projected/51a85cef-6a5b-4bde-b9fc-0685d2a96631-kube-api-access-k6q2h\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110774 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb0993f-dcb1-48a5-a415-9ce91bdf75ed-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zjwj9\" (UID: \"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110789 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-bound-sa-token\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110820 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb0993f-dcb1-48a5-a415-9ce91bdf75ed-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zjwj9\" (UID: \"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110844 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8dfr\" (UniqueName: \"kubernetes.io/projected/983bb1a5-259a-4d98-8c43-4d91d56825bb-kube-api-access-s8dfr\") pod \"package-server-manager-789f6589d5-wslzr\" (UID: \"983bb1a5-259a-4d98-8c43-4d91d56825bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110875 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-encryption-config\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.110891 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-proxy-tls\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.111456 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-console-serving-cert\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.111492 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-plugins-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.111507 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.111530 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-service-ca\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.111548 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-image-import-ca\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.111623 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6-config\") pod \"kube-controller-manager-operator-78b949d7b-j992c\" (UID: \"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.111639 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-audit\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.126902 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.132643 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.150955 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.214554 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.214643 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.215423 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:49.715408643 +0000 UTC m=+143.451121394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.223954 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/42491e41-75ad-4087-90ea-69ec8d173716-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2jr5f\" (UID: \"42491e41-75ad-4087-90ea-69ec8d173716\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.223988 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/42491e41-75ad-4087-90ea-69ec8d173716-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2jr5f\" (UID: \"42491e41-75ad-4087-90ea-69ec8d173716\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224052 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/73b144f3-32f0-47ca-96dc-54eab0196a50-metrics-tls\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224074 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b65e191f-43ce-4874-82b4-55cebaa7253b-certs\") pod \"machine-config-server-bkbbg\" (UID: \"b65e191f-43ce-4874-82b4-55cebaa7253b\") " pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224092 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b65e191f-43ce-4874-82b4-55cebaa7253b-node-bootstrap-token\") pod \"machine-config-server-bkbbg\" (UID: \"b65e191f-43ce-4874-82b4-55cebaa7253b\") " pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224113 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-tls\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224128 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-etcd-serving-ca\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224147 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-certificates\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224164 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224183 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6q2h\" (UniqueName: \"kubernetes.io/projected/51a85cef-6a5b-4bde-b9fc-0685d2a96631-kube-api-access-k6q2h\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224200 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5604a4e2-1b7b-40a6-9c33-38469fb706b4-config\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224221 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb0993f-dcb1-48a5-a415-9ce91bdf75ed-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zjwj9\" (UID: \"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224236 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px2ds\" (UniqueName: \"kubernetes.io/projected/d6fdb767-e022-4bdd-8464-753769e47a2a-kube-api-access-px2ds\") pod \"dns-default-wgtxp\" (UID: \"d6fdb767-e022-4bdd-8464-753769e47a2a\") " pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224252 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-bound-sa-token\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224282 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb0993f-dcb1-48a5-a415-9ce91bdf75ed-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zjwj9\" (UID: \"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224298 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8dfr\" (UniqueName: \"kubernetes.io/projected/983bb1a5-259a-4d98-8c43-4d91d56825bb-kube-api-access-s8dfr\") pod \"package-server-manager-789f6589d5-wslzr\" (UID: \"983bb1a5-259a-4d98-8c43-4d91d56825bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224317 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-encryption-config\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224334 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-proxy-tls\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224388 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-console-serving-cert\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224417 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-plugins-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224433 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224462 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-service-ca\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224479 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-image-import-ca\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224495 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6-config\") pod \"kube-controller-manager-operator-78b949d7b-j992c\" (UID: \"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224511 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-audit\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224530 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-serving-cert\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224546 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w65tp\" (UniqueName: \"kubernetes.io/projected/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-kube-api-access-w65tp\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224564 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224578 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73b144f3-32f0-47ca-96dc-54eab0196a50-trusted-ca\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224598 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224619 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5604a4e2-1b7b-40a6-9c33-38469fb706b4-etcd-ca\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224636 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6070dad-196d-4179-858c-4f75ddb71ebc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-66x5j\" (UID: \"b6070dad-196d-4179-858c-4f75ddb71ebc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224660 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhdtv\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-kube-api-access-jhdtv\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224675 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-oauth-serving-cert\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224693 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j992c\" (UID: \"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224725 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkvtw\" (UniqueName: \"kubernetes.io/projected/e4268b59-7563-4783-9957-2d1656b4ddec-kube-api-access-rkvtw\") pod \"ingress-canary-4h7ff\" (UID: \"e4268b59-7563-4783-9957-2d1656b4ddec\") " pod="openshift-ingress-canary/ingress-canary-4h7ff" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224745 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-csi-data-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224760 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/56c57ff0-baf4-483e-8763-a3ba52c0130e-tmpfs\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224777 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gxcpk\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224798 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7sng\" (UniqueName: \"kubernetes.io/projected/42491e41-75ad-4087-90ea-69ec8d173716-kube-api-access-c7sng\") pod \"multus-admission-controller-857f4d67dd-2jr5f\" (UID: \"42491e41-75ad-4087-90ea-69ec8d173716\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224824 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6070dad-196d-4179-858c-4f75ddb71ebc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-66x5j\" (UID: \"b6070dad-196d-4179-858c-4f75ddb71ebc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224841 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/56c57ff0-baf4-483e-8763-a3ba52c0130e-apiservice-cert\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224867 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/983bb1a5-259a-4d98-8c43-4d91d56825bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wslzr\" (UID: \"983bb1a5-259a-4d98-8c43-4d91d56825bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224884 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/56c57ff0-baf4-483e-8763-a3ba52c0130e-webhook-cert\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224899 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5wgb\" (UniqueName: \"kubernetes.io/projected/b6070dad-196d-4179-858c-4f75ddb71ebc-kube-api-access-m5wgb\") pod \"openshift-controller-manager-operator-756b6f6bc6-66x5j\" (UID: \"b6070dad-196d-4179-858c-4f75ddb71ebc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224918 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xwpp\" (UniqueName: \"kubernetes.io/projected/5604a4e2-1b7b-40a6-9c33-38469fb706b4-kube-api-access-8xwpp\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224937 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl6c4\" (UniqueName: \"kubernetes.io/projected/cfb0993f-dcb1-48a5-a415-9ce91bdf75ed-kube-api-access-jl6c4\") pod \"openshift-apiserver-operator-796bbdcf4f-zjwj9\" (UID: \"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224964 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-console-oauth-config\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.224979 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j992c\" (UID: \"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225000 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-mountpoint-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225014 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-node-pullsecrets\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225029 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5604a4e2-1b7b-40a6-9c33-38469fb706b4-etcd-service-ca\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225045 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gxcpk\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225063 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7vtf\" (UniqueName: \"kubernetes.io/projected/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-kube-api-access-d7vtf\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225079 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6fdb767-e022-4bdd-8464-753769e47a2a-metrics-tls\") pod \"dns-default-wgtxp\" (UID: \"d6fdb767-e022-4bdd-8464-753769e47a2a\") " pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225097 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac4bf154-96bf-4052-b8e6-371ede515470-config\") pod \"kube-apiserver-operator-766d6c64bb-xkbpk\" (UID: \"ac4bf154-96bf-4052-b8e6-371ede515470\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225113 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac4bf154-96bf-4052-b8e6-371ede515470-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xkbpk\" (UID: \"ac4bf154-96bf-4052-b8e6-371ede515470\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225128 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73b144f3-32f0-47ca-96dc-54eab0196a50-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225142 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p67k\" (UniqueName: \"kubernetes.io/projected/b0373e24-779c-47ec-b8d7-d4152b995948-kube-api-access-5p67k\") pod \"cluster-samples-operator-665b6dd947-p2b72\" (UID: \"b0373e24-779c-47ec-b8d7-d4152b995948\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225161 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-socket-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225175 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-registration-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225199 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225215 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-audit-dir\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225261 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-console-config\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225284 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-config\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225299 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggvrj\" (UniqueName: \"kubernetes.io/projected/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-kube-api-access-ggvrj\") pod \"marketplace-operator-79b997595-gxcpk\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225315 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z24n2\" (UniqueName: \"kubernetes.io/projected/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-kube-api-access-z24n2\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225331 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vksgw\" (UniqueName: \"kubernetes.io/projected/b65e191f-43ce-4874-82b4-55cebaa7253b-kube-api-access-vksgw\") pod \"machine-config-server-bkbbg\" (UID: \"b65e191f-43ce-4874-82b4-55cebaa7253b\") " pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225394 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktd5k\" (UniqueName: \"kubernetes.io/projected/73b144f3-32f0-47ca-96dc-54eab0196a50-kube-api-access-ktd5k\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225422 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e4268b59-7563-4783-9957-2d1656b4ddec-cert\") pod \"ingress-canary-4h7ff\" (UID: \"e4268b59-7563-4783-9957-2d1656b4ddec\") " pod="openshift-ingress-canary/ingress-canary-4h7ff" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225441 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5604a4e2-1b7b-40a6-9c33-38469fb706b4-etcd-client\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225458 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac4bf154-96bf-4052-b8e6-371ede515470-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xkbpk\" (UID: \"ac4bf154-96bf-4052-b8e6-371ede515470\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225481 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5604a4e2-1b7b-40a6-9c33-38469fb706b4-serving-cert\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225524 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg2f4\" (UniqueName: \"kubernetes.io/projected/56c57ff0-baf4-483e-8763-a3ba52c0130e-kube-api-access-fg2f4\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225543 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-trusted-ca\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225562 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6fdb767-e022-4bdd-8464-753769e47a2a-config-volume\") pod \"dns-default-wgtxp\" (UID: \"d6fdb767-e022-4bdd-8464-753769e47a2a\") " pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225579 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b0373e24-779c-47ec-b8d7-d4152b995948-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p2b72\" (UID: \"b0373e24-779c-47ec-b8d7-d4152b995948\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225595 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-etcd-client\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225618 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-images\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.225635 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-trusted-ca-bundle\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.228494 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-console-config\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.228869 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-socket-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.228933 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-registration-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.230045 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.230108 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-audit-dir\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.232522 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-image-import-ca\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.233265 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5604a4e2-1b7b-40a6-9c33-38469fb706b4-etcd-service-ca\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.233308 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-etcd-serving-ca\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.234302 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-mountpoint-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.234338 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-node-pullsecrets\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.234812 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6-config\") pod \"kube-controller-manager-operator-78b949d7b-j992c\" (UID: \"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.235196 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-audit\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.252430 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-trusted-ca-bundle\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.252572 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-service-ca\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.252575 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-serving-cert\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.252836 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-proxy-tls\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.253073 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j992c\" (UID: \"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.253770 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.253917 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-oauth-serving-cert\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.254541 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-csi-data-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.254834 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/56c57ff0-baf4-483e-8763-a3ba52c0130e-tmpfs\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.255001 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb0993f-dcb1-48a5-a415-9ce91bdf75ed-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zjwj9\" (UID: \"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.255255 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6070dad-196d-4179-858c-4f75ddb71ebc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-66x5j\" (UID: \"b6070dad-196d-4179-858c-4f75ddb71ebc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.235665 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-trusted-ca\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.255342 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.255962 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-encryption-config\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.256062 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-certificates\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.256549 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/983bb1a5-259a-4d98-8c43-4d91d56825bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wslzr\" (UID: \"983bb1a5-259a-4d98-8c43-4d91d56825bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.257098 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac4bf154-96bf-4052-b8e6-371ede515470-config\") pod \"kube-apiserver-operator-766d6c64bb-xkbpk\" (UID: \"ac4bf154-96bf-4052-b8e6-371ede515470\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.257218 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-tls\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.258115 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.258341 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/73b144f3-32f0-47ca-96dc-54eab0196a50-metrics-tls\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.259130 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-console-oauth-config\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.259600 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5604a4e2-1b7b-40a6-9c33-38469fb706b4-etcd-ca\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.259834 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:49.759816513 +0000 UTC m=+143.495529264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.262172 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.264099 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5604a4e2-1b7b-40a6-9c33-38469fb706b4-config\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.267173 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-images\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.267447 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6070dad-196d-4179-858c-4f75ddb71ebc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-66x5j\" (UID: \"b6070dad-196d-4179-858c-4f75ddb71ebc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.267955 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-config\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: W0217 00:07:49.268952 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedcf1adb_8ea8_4b8f_86c6_45d882e984c1.slice/crio-7c016e0046faed998d3fd6ee06af32891cd38e1fda974ac8e2deda5173a5a351 WatchSource:0}: Error finding container 7c016e0046faed998d3fd6ee06af32891cd38e1fda974ac8e2deda5173a5a351: Status 404 returned error can't find the container with id 7c016e0046faed998d3fd6ee06af32891cd38e1fda974ac8e2deda5173a5a351 Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.269644 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/56c57ff0-baf4-483e-8763-a3ba52c0130e-apiservice-cert\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.269892 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73b144f3-32f0-47ca-96dc-54eab0196a50-trusted-ca\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.270300 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/51a85cef-6a5b-4bde-b9fc-0685d2a96631-plugins-dir\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.270886 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gxcpk\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.272035 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e4268b59-7563-4783-9957-2d1656b4ddec-cert\") pod \"ingress-canary-4h7ff\" (UID: \"e4268b59-7563-4783-9957-2d1656b4ddec\") " pod="openshift-ingress-canary/ingress-canary-4h7ff" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.273426 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-console-serving-cert\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.275091 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb0993f-dcb1-48a5-a415-9ce91bdf75ed-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zjwj9\" (UID: \"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.282950 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac4bf154-96bf-4052-b8e6-371ede515470-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xkbpk\" (UID: \"ac4bf154-96bf-4052-b8e6-371ede515470\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.283021 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/56c57ff0-baf4-483e-8763-a3ba52c0130e-webhook-cert\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.285172 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-etcd-client\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.286098 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-bound-sa-token\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.287828 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gxcpk\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.290005 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5604a4e2-1b7b-40a6-9c33-38469fb706b4-serving-cert\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.290718 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5604a4e2-1b7b-40a6-9c33-38469fb706b4-etcd-client\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.291250 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.292312 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b0373e24-779c-47ec-b8d7-d4152b995948-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p2b72\" (UID: \"b0373e24-779c-47ec-b8d7-d4152b995948\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.306287 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8dfr\" (UniqueName: \"kubernetes.io/projected/983bb1a5-259a-4d98-8c43-4d91d56825bb-kube-api-access-s8dfr\") pod \"package-server-manager-789f6589d5-wslzr\" (UID: \"983bb1a5-259a-4d98-8c43-4d91d56825bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.310931 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac4bf154-96bf-4052-b8e6-371ede515470-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xkbpk\" (UID: \"ac4bf154-96bf-4052-b8e6-371ede515470\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.326672 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.326961 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:49.826918419 +0000 UTC m=+143.562631170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.327040 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6fdb767-e022-4bdd-8464-753769e47a2a-metrics-tls\") pod \"dns-default-wgtxp\" (UID: \"d6fdb767-e022-4bdd-8464-753769e47a2a\") " pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.327100 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vksgw\" (UniqueName: \"kubernetes.io/projected/b65e191f-43ce-4874-82b4-55cebaa7253b-kube-api-access-vksgw\") pod \"machine-config-server-bkbbg\" (UID: \"b65e191f-43ce-4874-82b4-55cebaa7253b\") " pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.327152 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6fdb767-e022-4bdd-8464-753769e47a2a-config-volume\") pod \"dns-default-wgtxp\" (UID: \"d6fdb767-e022-4bdd-8464-753769e47a2a\") " pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.327316 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b65e191f-43ce-4874-82b4-55cebaa7253b-certs\") pod \"machine-config-server-bkbbg\" (UID: \"b65e191f-43ce-4874-82b4-55cebaa7253b\") " pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.327337 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b65e191f-43ce-4874-82b4-55cebaa7253b-node-bootstrap-token\") pod \"machine-config-server-bkbbg\" (UID: \"b65e191f-43ce-4874-82b4-55cebaa7253b\") " pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.328021 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px2ds\" (UniqueName: \"kubernetes.io/projected/d6fdb767-e022-4bdd-8464-753769e47a2a-kube-api-access-px2ds\") pod \"dns-default-wgtxp\" (UID: \"d6fdb767-e022-4bdd-8464-753769e47a2a\") " pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.327860 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73b144f3-32f0-47ca-96dc-54eab0196a50-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.328056 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6fdb767-e022-4bdd-8464-753769e47a2a-config-volume\") pod \"dns-default-wgtxp\" (UID: \"d6fdb767-e022-4bdd-8464-753769e47a2a\") " pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.328070 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.328574 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:49.828345001 +0000 UTC m=+143.564057752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.331347 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b65e191f-43ce-4874-82b4-55cebaa7253b-node-bootstrap-token\") pod \"machine-config-server-bkbbg\" (UID: \"b65e191f-43ce-4874-82b4-55cebaa7253b\") " pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.331831 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6fdb767-e022-4bdd-8464-753769e47a2a-metrics-tls\") pod \"dns-default-wgtxp\" (UID: \"d6fdb767-e022-4bdd-8464-753769e47a2a\") " pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.334346 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b65e191f-43ce-4874-82b4-55cebaa7253b-certs\") pod \"machine-config-server-bkbbg\" (UID: \"b65e191f-43ce-4874-82b4-55cebaa7253b\") " pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.346621 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-c6pc7"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.347981 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.349342 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p67k\" (UniqueName: \"kubernetes.io/projected/b0373e24-779c-47ec-b8d7-d4152b995948-kube-api-access-5p67k\") pod \"cluster-samples-operator-665b6dd947-p2b72\" (UID: \"b0373e24-779c-47ec-b8d7-d4152b995948\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.366771 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhdtv\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-kube-api-access-jhdtv\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.395988 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6q2h\" (UniqueName: \"kubernetes.io/projected/51a85cef-6a5b-4bde-b9fc-0685d2a96631-kube-api-access-k6q2h\") pod \"csi-hostpathplugin-j28sv\" (UID: \"51a85cef-6a5b-4bde-b9fc-0685d2a96631\") " pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.424710 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xwpp\" (UniqueName: \"kubernetes.io/projected/5604a4e2-1b7b-40a6-9c33-38469fb706b4-kube-api-access-8xwpp\") pod \"etcd-operator-b45778765-g8jrf\" (UID: \"5604a4e2-1b7b-40a6-9c33-38469fb706b4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.429120 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.429576 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:49.929560171 +0000 UTC m=+143.665272922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.437790 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl6c4\" (UniqueName: \"kubernetes.io/projected/cfb0993f-dcb1-48a5-a415-9ce91bdf75ed-kube-api-access-jl6c4\") pod \"openshift-apiserver-operator-796bbdcf4f-zjwj9\" (UID: \"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.449134 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkvtw\" (UniqueName: \"kubernetes.io/projected/e4268b59-7563-4783-9957-2d1656b4ddec-kube-api-access-rkvtw\") pod \"ingress-canary-4h7ff\" (UID: \"e4268b59-7563-4783-9957-2d1656b4ddec\") " pod="openshift-ingress-canary/ingress-canary-4h7ff" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.476102 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7sng\" (UniqueName: \"kubernetes.io/projected/42491e41-75ad-4087-90ea-69ec8d173716-kube-api-access-c7sng\") pod \"multus-admission-controller-857f4d67dd-2jr5f\" (UID: \"42491e41-75ad-4087-90ea-69ec8d173716\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.492614 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7vtf\" (UniqueName: \"kubernetes.io/projected/8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7-kube-api-access-d7vtf\") pod \"console-f9d7485db-687jq\" (UID: \"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7\") " pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.511531 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktd5k\" (UniqueName: \"kubernetes.io/projected/73b144f3-32f0-47ca-96dc-54eab0196a50-kube-api-access-ktd5k\") pod \"ingress-operator-5b745b69d9-4pvmr\" (UID: \"73b144f3-32f0-47ca-96dc-54eab0196a50\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.517528 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.525991 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.533709 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.535330 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.535610 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.035597825 +0000 UTC m=+143.771310576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.538600 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg2f4\" (UniqueName: \"kubernetes.io/projected/56c57ff0-baf4-483e-8763-a3ba52c0130e-kube-api-access-fg2f4\") pod \"packageserver-d55dfcdfc-2tt5n\" (UID: \"56c57ff0-baf4-483e-8763-a3ba52c0130e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.540292 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.559280 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.579635 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.579987 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h9fqv"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.582129 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w65tp\" (UniqueName: \"kubernetes.io/projected/3db9a5a7-2885-4a36-9a69-a38957ddb5a3-kube-api-access-w65tp\") pod \"apiserver-76f77b778f-rdw2m\" (UID: \"3db9a5a7-2885-4a36-9a69-a38957ddb5a3\") " pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.588747 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5wgb\" (UniqueName: \"kubernetes.io/projected/b6070dad-196d-4179-858c-4f75ddb71ebc-kube-api-access-m5wgb\") pod \"openshift-controller-manager-operator-756b6f6bc6-66x5j\" (UID: \"b6070dad-196d-4179-858c-4f75ddb71ebc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.594780 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.596871 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j992c\" (UID: \"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.602870 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.606274 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggvrj\" (UniqueName: \"kubernetes.io/projected/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-kube-api-access-ggvrj\") pod \"marketplace-operator-79b997595-gxcpk\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.610440 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.627812 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z24n2\" (UniqueName: \"kubernetes.io/projected/7c9dd9bc-8d10-4946-a664-d7e5d754c88d-kube-api-access-z24n2\") pod \"machine-config-operator-74547568cd-n8hnq\" (UID: \"7c9dd9bc-8d10-4946-a664-d7e5d754c88d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.631860 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.635642 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.635989 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.1359751 +0000 UTC m=+143.871687851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.649926 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4h7ff" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.670284 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vksgw\" (UniqueName: \"kubernetes.io/projected/b65e191f-43ce-4874-82b4-55cebaa7253b-kube-api-access-vksgw\") pod \"machine-config-server-bkbbg\" (UID: \"b65e191f-43ce-4874-82b4-55cebaa7253b\") " pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.678156 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-j28sv" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.687650 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px2ds\" (UniqueName: \"kubernetes.io/projected/d6fdb767-e022-4bdd-8464-753769e47a2a-kube-api-access-px2ds\") pod \"dns-default-wgtxp\" (UID: \"d6fdb767-e022-4bdd-8464-753769e47a2a\") " pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.695424 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.712177 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bkbbg" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.737073 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.737509 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.237495009 +0000 UTC m=+143.973207760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.790844 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.794637 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.806542 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29521440-6zdhx"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.808422 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.831375 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-glck9" event={"ID":"f56490e5-1c42-4126-b7f5-cd7ef73b32bd","Type":"ContainerStarted","Data":"fa8b957c25bb16fa54cde9ac7daf44c6e6358322f45a5ed390e53f7765be86cc"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.831423 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-glck9" event={"ID":"f56490e5-1c42-4126-b7f5-cd7ef73b32bd","Type":"ContainerStarted","Data":"bad69b9d3ca6e3f9e7aadd8810cbeef6b1024c47794ae1ff0d8801930ced4a98"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.834381 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" event={"ID":"0936c3dd-a635-4721-8499-cd99755c8de8","Type":"ContainerStarted","Data":"688647415e2140735a7218a020ed18614a5d64b38923ca5f26110f9b67447d5e"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.834410 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" event={"ID":"0936c3dd-a635-4721-8499-cd99755c8de8","Type":"ContainerStarted","Data":"6e511932171519cc772cf22ee2541eb5364e032d06cfde2bbe457a88b757e54b"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.834421 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" event={"ID":"0936c3dd-a635-4721-8499-cd99755c8de8","Type":"ContainerStarted","Data":"1c0baf71ec68e84bb67003274b0ebad19104ef2ef97d8fff855efcd0023a12f2"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.840153 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" event={"ID":"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c","Type":"ContainerStarted","Data":"6c5b88baddb81569224b0813f534f1a556f0fb2f7a9ebac30fabf9eacf1489d2"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.841583 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" event={"ID":"e069c9f2-7d28-4499-b432-c1bd329fcd3a","Type":"ContainerStarted","Data":"ebadef7d66c49f0987fa8489c78fcdd62715055f8df0b4c7e3473b5532ba87aa"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.841608 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" event={"ID":"e069c9f2-7d28-4499-b432-c1bd329fcd3a","Type":"ContainerStarted","Data":"f030092c04ad12ec47bb84612d387a95db437a644ce1ae6638da10673e4c2fc5"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.842491 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.842873 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.843284 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.343269485 +0000 UTC m=+144.078982236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.849650 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" event={"ID":"edcf1adb-8ea8-4b8f-86c6-45d882e984c1","Type":"ContainerStarted","Data":"50ecce2d023f7ae6235c0bfed8c5fe19e327262622814889d4ac79be2661ae37"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.849696 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" event={"ID":"edcf1adb-8ea8-4b8f-86c6-45d882e984c1","Type":"ContainerStarted","Data":"7c016e0046faed998d3fd6ee06af32891cd38e1fda974ac8e2deda5173a5a351"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.850562 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.855679 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" event={"ID":"8daa0570-a319-44d6-be28-c5717cb07575","Type":"ContainerStarted","Data":"47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.855884 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" event={"ID":"8daa0570-a319-44d6-be28-c5717cb07575","Type":"ContainerStarted","Data":"2a395ebd5fa6c242d817997f09e61db1fa0daf7dc0880751e34f5a32373d8350"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.856228 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.856671 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.865746 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.885997 4690 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-d67hq container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.886087 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" podUID="e069c9f2-7d28-4499-b432-c1bd329fcd3a" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.886116 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.889114 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" event={"ID":"159fd3d3-b560-4f05-b24e-25611073d6b8","Type":"ContainerStarted","Data":"4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.889152 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" event={"ID":"159fd3d3-b560-4f05-b24e-25611073d6b8","Type":"ContainerStarted","Data":"baac3fc17b094768715a02a9e059608c769d8977b936d108b1095394d167cae3"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.889493 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.891451 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" event={"ID":"b7b8197f-4d82-4a0b-bed6-485dfaab9024","Type":"ContainerStarted","Data":"dc33dc5fc7af820e3e7edc572c362cba708bcf944b49af38c0a57cd91cd475c2"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.893814 4690 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-wkxlb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.893858 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" podUID="edcf1adb-8ea8-4b8f-86c6-45d882e984c1" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.893811 4690 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-gzx5v container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.893914 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" podUID="8daa0570-a319-44d6-be28-c5717cb07575" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.894939 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" event={"ID":"8601b0b5-9982-4e47-98b5-f36b9ceb89ff","Type":"ContainerStarted","Data":"d9aa8d1fab38a0312ec02dc4604646bc22c382c401ae9c61223a06c389303468"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.895001 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" event={"ID":"8601b0b5-9982-4e47-98b5-f36b9ceb89ff","Type":"ContainerStarted","Data":"51b24c3637ee8ca09f3d0a606b0b3352643ee5c4734f8cf8c9f6f364d47407e4"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.896555 4690 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-zw657 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.896607 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" podUID="159fd3d3-b560-4f05-b24e-25611073d6b8" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.897474 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l" event={"ID":"50b867d9-7f2c-44b0-a432-95ae7e82f673","Type":"ContainerStarted","Data":"fae721a1db03b4cabfc74388835f3dfe8080f4cb32943648906390b593447c31"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.897500 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l" event={"ID":"50b867d9-7f2c-44b0-a432-95ae7e82f673","Type":"ContainerStarted","Data":"07f28a4ddff36b1563e8edff09f333a6acc93fce38109f4f9a257c56b0442930"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.901602 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" event={"ID":"0903ddf6-73c1-4f80-99de-54b204d9df38","Type":"ContainerStarted","Data":"9d5ea9864e4a56b9bb9274154cd14227539c8b86c8e90f56299d90b8bcae7472"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.901641 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" event={"ID":"0903ddf6-73c1-4f80-99de-54b204d9df38","Type":"ContainerStarted","Data":"f921a8d8ad9b559bd659da0395e11e7994029b3ba31a599e02a7af523e012a33"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.918674 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.936000 4690 generic.go:334] "Generic (PLEG): container finished" podID="e265f262-30e7-430d-b0f7-05db8db0c03d" containerID="b7389813dc98fa90ebebf49631b81992753d2c364ff2e7f99417d4a8d173e8d2" exitCode=0 Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.936045 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" event={"ID":"e265f262-30e7-430d-b0f7-05db8db0c03d","Type":"ContainerDied","Data":"b7389813dc98fa90ebebf49631b81992753d2c364ff2e7f99417d4a8d173e8d2"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.936070 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" event={"ID":"e265f262-30e7-430d-b0f7-05db8db0c03d","Type":"ContainerStarted","Data":"0211998d97d9d10163920b5ae5a60d647421650dc021d3c7f64fab1e522a1886"} Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.945152 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:49 crc kubenswrapper[4690]: E0217 00:07:49.949222 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.449209375 +0000 UTC m=+144.184922126 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.955264 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps"] Feb 17 00:07:49 crc kubenswrapper[4690]: I0217 00:07:49.963618 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xflvn"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.039136 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.045590 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vwd2q"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.045875 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.046143 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.546126198 +0000 UTC m=+144.281838949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.046563 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.048329 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.548312653 +0000 UTC m=+144.284025404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.096227 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.098032 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.098080 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.148638 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.149020 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.649004787 +0000 UTC m=+144.384717538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.226612 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.246676 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pbbhc"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.250421 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.260034 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.260474 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.760463782 +0000 UTC m=+144.496176533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.318234 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.345009 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-glck9" podStartSLOduration=123.344995556 podStartE2EDuration="2m3.344995556s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:50.343229714 +0000 UTC m=+144.078942475" watchObservedRunningTime="2026-02-17 00:07:50.344995556 +0000 UTC m=+144.080708307" Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.348954 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.357458 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-687jq"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.361193 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.361444 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.861423555 +0000 UTC m=+144.597136306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.361546 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.362008 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.861994082 +0000 UTC m=+144.597706833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.375037 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.393102 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr"] Feb 17 00:07:50 crc kubenswrapper[4690]: W0217 00:07:50.415831 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51b6988f_45a7_4faa_9f25_0c03e4fe316b.slice/crio-acc5628bc365e630f6e4129a59814d9f7e5a54bd318e619fe4ce00ddcc9370d6 WatchSource:0}: Error finding container acc5628bc365e630f6e4129a59814d9f7e5a54bd318e619fe4ce00ddcc9370d6: Status 404 returned error can't find the container with id acc5628bc365e630f6e4129a59814d9f7e5a54bd318e619fe4ce00ddcc9370d6 Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.432125 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" podStartSLOduration=122.432105797 podStartE2EDuration="2m2.432105797s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:50.431466698 +0000 UTC m=+144.167179449" watchObservedRunningTime="2026-02-17 00:07:50.432105797 +0000 UTC m=+144.167818548" Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.463379 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.463747 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.963720277 +0000 UTC m=+144.699433038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.463956 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.464290 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:50.964275653 +0000 UTC m=+144.699988404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.468924 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-j28sv"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.539986 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.544586 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-g8jrf"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.545170 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.565890 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.566184 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:51.066169394 +0000 UTC m=+144.801882145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.638775 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-vpd8p" podStartSLOduration=123.638756822 podStartE2EDuration="2m3.638756822s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:50.63867675 +0000 UTC m=+144.374389501" watchObservedRunningTime="2026-02-17 00:07:50.638756822 +0000 UTC m=+144.374469573" Feb 17 00:07:50 crc kubenswrapper[4690]: W0217 00:07:50.642393 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56c57ff0_baf4_483e_8763_a3ba52c0130e.slice/crio-41a782415d835c7654c5d8b605421dac3cb4d4f1918ee6b0461a3945afd10122 WatchSource:0}: Error finding container 41a782415d835c7654c5d8b605421dac3cb4d4f1918ee6b0461a3945afd10122: Status 404 returned error can't find the container with id 41a782415d835c7654c5d8b605421dac3cb4d4f1918ee6b0461a3945afd10122 Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.676284 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.676611 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:51.176600588 +0000 UTC m=+144.912313339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.777236 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.777823 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:51.277808558 +0000 UTC m=+145.013521309 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.784903 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4h7ff"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.784947 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gxcpk"] Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.879185 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.879462 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:51.37944984 +0000 UTC m=+145.115162591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.968180 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" event={"ID":"acc970de-869b-4a13-8d0e-94370922b22f","Type":"ContainerStarted","Data":"143f5fc99d4f44f61fdc565790c4a7c6110213640ba642a9750f2ab1f2e139a6"} Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.980347 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:50 crc kubenswrapper[4690]: E0217 00:07:50.980869 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:51.480846185 +0000 UTC m=+145.216558976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:50 crc kubenswrapper[4690]: I0217 00:07:50.996150 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" podStartSLOduration=122.99613678 podStartE2EDuration="2m2.99613678s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:50.995776849 +0000 UTC m=+144.731489600" watchObservedRunningTime="2026-02-17 00:07:50.99613678 +0000 UTC m=+144.731849531" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.011191 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29521440-6zdhx" event={"ID":"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6","Type":"ContainerStarted","Data":"c5c229e4667e388858094b7cf9f5cc1ae1358905c1be1fb0a8289738a65fd4b4"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.018071 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" event={"ID":"8601b0b5-9982-4e47-98b5-f36b9ceb89ff","Type":"ContainerStarted","Data":"f0865183b8dd9f3da15bf2b24b8f6e44936e41b8f5c20f2eccecbf6d83a13d8f"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.048704 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" event={"ID":"5604a4e2-1b7b-40a6-9c33-38469fb706b4","Type":"ContainerStarted","Data":"2778c058b36142a0ef45a25964c2bfa84021a92e0ce2d0b5c7acd4c70059e037"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.056706 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" event={"ID":"0ded4d5d-1f0b-4caf-a498-73f7963446fb","Type":"ContainerStarted","Data":"5fc7174276c0a442867c7eead98715c552b800025879cbaf9056c96178094a2d"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.065734 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq"] Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.067147 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wgtxp"] Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.067222 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" podStartSLOduration=123.067202113 podStartE2EDuration="2m3.067202113s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.033505301 +0000 UTC m=+144.769218052" watchObservedRunningTime="2026-02-17 00:07:51.067202113 +0000 UTC m=+144.802914864" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.074983 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2jr5f"] Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.098676 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:51 crc kubenswrapper[4690]: E0217 00:07:51.099032 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:51.599018149 +0000 UTC m=+145.334730900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.099576 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" event={"ID":"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed","Type":"ContainerStarted","Data":"fe2c2f03d32ac145b54f783f9753233aabdeb315af09bf38fdad2631fa26c04a"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.104103 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:07:51 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:07:51 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:07:51 crc kubenswrapper[4690]: healthz check failed Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.104148 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:07:51 crc kubenswrapper[4690]: W0217 00:07:51.110955 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c9dd9bc_8d10_4946_a664_d7e5d754c88d.slice/crio-569f90688d724c47b963cf7be9b6529bc9b0b603a08906ed12f17054e77f3113 WatchSource:0}: Error finding container 569f90688d724c47b963cf7be9b6529bc9b0b603a08906ed12f17054e77f3113: Status 404 returned error can't find the container with id 569f90688d724c47b963cf7be9b6529bc9b0b603a08906ed12f17054e77f3113 Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.126251 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rdw2m"] Feb 17 00:07:51 crc kubenswrapper[4690]: W0217 00:07:51.131331 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42491e41_75ad_4087_90ea_69ec8d173716.slice/crio-47a34be16e0a7503d4b68af624c5206a786543e898fedb009722b06c42bb8933 WatchSource:0}: Error finding container 47a34be16e0a7503d4b68af624c5206a786543e898fedb009722b06c42bb8933: Status 404 returned error can't find the container with id 47a34be16e0a7503d4b68af624c5206a786543e898fedb009722b06c42bb8933 Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.151001 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" event={"ID":"21ad60ce-9b8d-4dfc-bc60-06901e688248","Type":"ContainerStarted","Data":"e7515a466bfd584184ceceb03a2e526889c6ddcc92569a8059e456f6fdad46e5"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.163979 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" event={"ID":"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c","Type":"ContainerStarted","Data":"81f2927e009458c5460d2a405d1134cd47d6306812010b635bbb5f903a9ef665"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.164930 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j"] Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.183012 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" event={"ID":"b85e9f0e-5e1f-4d99-967e-1a3d6b12202c","Type":"ContainerStarted","Data":"9f2ebca6d2fd93030bfa878509a9edb02fcf7d4a4fe57bcea8f708bd2998abd8"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.193917 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" podStartSLOduration=124.193899671 podStartE2EDuration="2m4.193899671s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.19216827 +0000 UTC m=+144.927881021" watchObservedRunningTime="2026-02-17 00:07:51.193899671 +0000 UTC m=+144.929612422" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.205904 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:51 crc kubenswrapper[4690]: E0217 00:07:51.207063 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:51.707049062 +0000 UTC m=+145.442761803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.207167 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" event={"ID":"51b6988f-45a7-4faa-9f25-0c03e4fe316b","Type":"ContainerStarted","Data":"acc5628bc365e630f6e4129a59814d9f7e5a54bd318e619fe4ce00ddcc9370d6"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.248862 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l" event={"ID":"50b867d9-7f2c-44b0-a432-95ae7e82f673","Type":"ContainerStarted","Data":"b5f15b2de98258ab8f44873655e701c46dd8e573dc41f6691411bfbbdf9f023f"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.257623 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c"] Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.263946 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" event={"ID":"0903ddf6-73c1-4f80-99de-54b204d9df38","Type":"ContainerStarted","Data":"92e2453880c8e485cd33e4bead823baa51584396ee2a0c320f14617803743bdb"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.288514 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bkbbg" event={"ID":"b65e191f-43ce-4874-82b4-55cebaa7253b","Type":"ContainerStarted","Data":"f9a971e62fc51d05929f9036a862431fd4d752a60ad9b45e2af53cb554f43082"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.303124 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" event={"ID":"56c57ff0-baf4-483e-8763-a3ba52c0130e","Type":"ContainerStarted","Data":"41a782415d835c7654c5d8b605421dac3cb4d4f1918ee6b0461a3945afd10122"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.306803 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:51 crc kubenswrapper[4690]: E0217 00:07:51.308098 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:51.808087887 +0000 UTC m=+145.543800638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.310649 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" event={"ID":"ac4bf154-96bf-4052-b8e6-371ede515470","Type":"ContainerStarted","Data":"aa7eceff8d269ce75071989f67654861324708eca2d53a19909e203b88c45943"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.319403 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-687jq" event={"ID":"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7","Type":"ContainerStarted","Data":"a6a9b13910c3839363877e68045034916bfbbdcdbec70fe8572f4ca848670841"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.325414 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" event={"ID":"b7b8197f-4d82-4a0b-bed6-485dfaab9024","Type":"ContainerStarted","Data":"4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.325785 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.326861 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" event={"ID":"cd0a26f5-e912-40a5-a41f-3754297b6f68","Type":"ContainerStarted","Data":"199374e23126165a5a69dd695ee9dbfa088718c6d45783cb0e830745b7e5957b"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.340601 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" event={"ID":"a94ddfaa-d52b-4ded-ad8b-c0732e08a983","Type":"ContainerStarted","Data":"a1cc4fe4a104489a4c9c688307183ccdd9bfeb5cc5e7f1388a2667aae0336631"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.352005 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j28sv" event={"ID":"51a85cef-6a5b-4bde-b9fc-0685d2a96631","Type":"ContainerStarted","Data":"0a3acc67de460925669781faaeef30e760772623dd3896b2bb0b4852275169ae"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.363197 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" event={"ID":"0d6d4f41-49a8-4ddd-acfb-1695db85409c","Type":"ContainerStarted","Data":"8af25c56f41de60a150e3368f9202e974ac80fbbc1fb0a11d7337cdd9b99400c"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.373759 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pbbhc" event={"ID":"f06f4ec3-86a3-429b-8920-03dd8e7a4569","Type":"ContainerStarted","Data":"9fb972febc26618e6ce9d27908bba4c1babcd3e1788813e62fa4da8811b7d233"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.374310 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.376838 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4h7ff" event={"ID":"e4268b59-7563-4783-9957-2d1656b4ddec","Type":"ContainerStarted","Data":"682f3e2ea9170efa47a4de7491423ad3e8a0261d3fad0d0e2844734d6242bb66"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.383848 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" event={"ID":"e265f262-30e7-430d-b0f7-05db8db0c03d","Type":"ContainerStarted","Data":"97e3f3770556e95be0257a438d6897dfcf79506034e625375a1448b4a812a455"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.384467 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.385893 4690 patch_prober.go:28] interesting pod/console-operator-58897d9998-pbbhc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.385933 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-pbbhc" podUID="f06f4ec3-86a3-429b-8920-03dd8e7a4569" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.408432 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.411714 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" event={"ID":"ab15103a-6de2-47f7-8640-6c3253d2ee68","Type":"ContainerStarted","Data":"8c070f31839beb641dd83487b0857d4cdc3ce3d76c1718b310b1d8d865d3fd04"} Feb 17 00:07:51 crc kubenswrapper[4690]: E0217 00:07:51.413104 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:51.913080569 +0000 UTC m=+145.648793320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.425334 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-vgzbg" podStartSLOduration=123.425317373 podStartE2EDuration="2m3.425317373s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.423783178 +0000 UTC m=+145.159495929" watchObservedRunningTime="2026-02-17 00:07:51.425317373 +0000 UTC m=+145.161030124" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.426189 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xflvn" event={"ID":"929321b2-58ac-4633-af87-da506a21c23a","Type":"ContainerStarted","Data":"a92ce46861c48a1919b1c3dc36d1ad980f564661abcfa188bb579df29dafee4c"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.426228 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xflvn" event={"ID":"929321b2-58ac-4633-af87-da506a21c23a","Type":"ContainerStarted","Data":"2a8869f21e763c035341ea38a9e72c6200537378242f45088eb94972f043fdb3"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.429090 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-xflvn" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.435893 4690 patch_prober.go:28] interesting pod/downloads-7954f5f757-xflvn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.435944 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xflvn" podUID="929321b2-58ac-4633-af87-da506a21c23a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.443996 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" event={"ID":"983bb1a5-259a-4d98-8c43-4d91d56825bb","Type":"ContainerStarted","Data":"4b3e10492b93269e37f50d0e3573ac00f18801e17d320f13fd1895c9feea37de"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.450596 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" event={"ID":"73b144f3-32f0-47ca-96dc-54eab0196a50","Type":"ContainerStarted","Data":"3b0796d0c20821e0433361dc938c29950d418309c589130bb2200c1c26bc525f"} Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.456969 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.457102 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.474687 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d67hq" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.510345 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.512694 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wkxlb" Feb 17 00:07:51 crc kubenswrapper[4690]: E0217 00:07:51.513114 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.013098244 +0000 UTC m=+145.748810995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.522449 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-pbbhc" podStartSLOduration=124.522436022 podStartE2EDuration="2m4.522436022s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.520815043 +0000 UTC m=+145.256527794" watchObservedRunningTime="2026-02-17 00:07:51.522436022 +0000 UTC m=+145.258148773" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.576865 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" podStartSLOduration=124.5768482 podStartE2EDuration="2m4.5768482s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.574269183 +0000 UTC m=+145.309981934" watchObservedRunningTime="2026-02-17 00:07:51.5768482 +0000 UTC m=+145.312560941" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.616793 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:51 crc kubenswrapper[4690]: E0217 00:07:51.618080 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.118065956 +0000 UTC m=+145.853778707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.621617 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pld4l" podStartSLOduration=123.621601381 podStartE2EDuration="2m3.621601381s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.619965012 +0000 UTC m=+145.355677763" watchObservedRunningTime="2026-02-17 00:07:51.621601381 +0000 UTC m=+145.357314132" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.718713 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:51 crc kubenswrapper[4690]: E0217 00:07:51.718977 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.218967506 +0000 UTC m=+145.954680257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.734218 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" podStartSLOduration=124.734200119 podStartE2EDuration="2m4.734200119s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.689019976 +0000 UTC m=+145.424732727" watchObservedRunningTime="2026-02-17 00:07:51.734200119 +0000 UTC m=+145.469912870" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.763320 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" podStartSLOduration=124.763303245 podStartE2EDuration="2m4.763303245s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.73388683 +0000 UTC m=+145.469599581" watchObservedRunningTime="2026-02-17 00:07:51.763303245 +0000 UTC m=+145.499015986" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.821942 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:51 crc kubenswrapper[4690]: E0217 00:07:51.822109 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.322085053 +0000 UTC m=+146.057797804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.822233 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:51 crc kubenswrapper[4690]: E0217 00:07:51.822613 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.322605479 +0000 UTC m=+146.058318230 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.868430 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9jxf8" podStartSLOduration=123.868409941 podStartE2EDuration="2m3.868409941s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.851682913 +0000 UTC m=+145.587395664" watchObservedRunningTime="2026-02-17 00:07:51.868409941 +0000 UTC m=+145.604122692" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.875004 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" podStartSLOduration=123.874982436 podStartE2EDuration="2m3.874982436s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.79809828 +0000 UTC m=+145.533811031" watchObservedRunningTime="2026-02-17 00:07:51.874982436 +0000 UTC m=+145.610695187" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.895257 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bkbbg" podStartSLOduration=5.895241349 podStartE2EDuration="5.895241349s" podCreationTimestamp="2026-02-17 00:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.893099175 +0000 UTC m=+145.628811916" watchObservedRunningTime="2026-02-17 00:07:51.895241349 +0000 UTC m=+145.630954100" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.925792 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-h9fqv" podStartSLOduration=123.925767266 podStartE2EDuration="2m3.925767266s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:51.92218799 +0000 UTC m=+145.657900751" watchObservedRunningTime="2026-02-17 00:07:51.925767266 +0000 UTC m=+145.661480017" Feb 17 00:07:51 crc kubenswrapper[4690]: I0217 00:07:51.937084 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:51 crc kubenswrapper[4690]: E0217 00:07:51.938031 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.438014131 +0000 UTC m=+146.173726882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.038726 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:52 crc kubenswrapper[4690]: E0217 00:07:52.039521 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.539511119 +0000 UTC m=+146.275223870 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.108862 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:07:52 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:07:52 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:07:52 crc kubenswrapper[4690]: healthz check failed Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.108939 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.128561 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bdbdq" podStartSLOduration=125.128547247 podStartE2EDuration="2m5.128547247s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:52.127539547 +0000 UTC m=+145.863252298" watchObservedRunningTime="2026-02-17 00:07:52.128547247 +0000 UTC m=+145.864259988" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.139002 4690 csr.go:261] certificate signing request csr-j7jf4 is approved, waiting to be issued Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.147250 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:52 crc kubenswrapper[4690]: E0217 00:07:52.147553 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.647538082 +0000 UTC m=+146.383250833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.153662 4690 csr.go:257] certificate signing request csr-j7jf4 is issued Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.177210 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" podStartSLOduration=124.177192414 podStartE2EDuration="2m4.177192414s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:52.16495261 +0000 UTC m=+145.900665361" watchObservedRunningTime="2026-02-17 00:07:52.177192414 +0000 UTC m=+145.912905165" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.188742 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fcmdk"] Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.189811 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.197578 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fcmdk"] Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.201496 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.212847 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" podStartSLOduration=125.212832064 podStartE2EDuration="2m5.212832064s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:52.212251896 +0000 UTC m=+145.947964657" watchObservedRunningTime="2026-02-17 00:07:52.212832064 +0000 UTC m=+145.948544815" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.248754 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:52 crc kubenswrapper[4690]: E0217 00:07:52.249125 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.749109052 +0000 UTC m=+146.484821803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.326877 4690 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-c6pc7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.326936 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" podUID="b7b8197f-4d82-4a0b-bed6-485dfaab9024" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.348340 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-xflvn" podStartSLOduration=125.348306542 podStartE2EDuration="2m5.348306542s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:52.347606412 +0000 UTC m=+146.083319163" watchObservedRunningTime="2026-02-17 00:07:52.348306542 +0000 UTC m=+146.084019303" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.349656 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.349892 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kcw7\" (UniqueName: \"kubernetes.io/projected/4c5d01e4-052c-4541-a385-79a8066bba8f-kube-api-access-7kcw7\") pod \"community-operators-fcmdk\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.349951 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-utilities\") pod \"community-operators-fcmdk\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.349997 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-catalog-content\") pod \"community-operators-fcmdk\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: E0217 00:07:52.350095 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.850082875 +0000 UTC m=+146.585795626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.360313 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2tkvq"] Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.363214 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.373369 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.403207 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2tkvq"] Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.458178 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-catalog-content\") pod \"community-operators-fcmdk\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.458218 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-utilities\") pod \"certified-operators-2tkvq\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.458271 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.458301 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kcw7\" (UniqueName: \"kubernetes.io/projected/4c5d01e4-052c-4541-a385-79a8066bba8f-kube-api-access-7kcw7\") pod \"community-operators-fcmdk\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.458336 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2nz2\" (UniqueName: \"kubernetes.io/projected/c08acd43-80f4-4a48-b848-7899561e26b5-kube-api-access-b2nz2\") pod \"certified-operators-2tkvq\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.458404 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-utilities\") pod \"community-operators-fcmdk\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.458446 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-catalog-content\") pod \"certified-operators-2tkvq\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.458837 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-catalog-content\") pod \"community-operators-fcmdk\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: E0217 00:07:52.459087 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:52.959075937 +0000 UTC m=+146.694788688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.459708 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-utilities\") pod \"community-operators-fcmdk\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.495049 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kcw7\" (UniqueName: \"kubernetes.io/projected/4c5d01e4-052c-4541-a385-79a8066bba8f-kube-api-access-7kcw7\") pod \"community-operators-fcmdk\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.509841 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" event={"ID":"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6","Type":"ContainerStarted","Data":"989273e4b548cadf53732e8ef4cd755b2aeee103c1a910ae3aef7b7d05fcbd99"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.530563 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.542779 4690 generic.go:334] "Generic (PLEG): container finished" podID="cd0a26f5-e912-40a5-a41f-3754297b6f68" containerID="d0f6f1712042d9ee2652c749c260d6aa88005a2aadb723e43983ab9ff885f1c4" exitCode=0 Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.542891 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" event={"ID":"cd0a26f5-e912-40a5-a41f-3754297b6f68","Type":"ContainerDied","Data":"d0f6f1712042d9ee2652c749c260d6aa88005a2aadb723e43983ab9ff885f1c4"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.559101 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.559313 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-utilities\") pod \"certified-operators-2tkvq\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.559426 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2nz2\" (UniqueName: \"kubernetes.io/projected/c08acd43-80f4-4a48-b848-7899561e26b5-kube-api-access-b2nz2\") pod \"certified-operators-2tkvq\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.559504 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-catalog-content\") pod \"certified-operators-2tkvq\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.559976 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-catalog-content\") pod \"certified-operators-2tkvq\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: E0217 00:07:52.560057 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:53.060039889 +0000 UTC m=+146.795752640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.564792 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-utilities\") pod \"certified-operators-2tkvq\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.599446 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k62p2"] Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.615676 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" event={"ID":"cfb0993f-dcb1-48a5-a415-9ce91bdf75ed","Type":"ContainerStarted","Data":"402f691b00f63f1becc8b118dcfa232cf7d9d7bc8c0d9158d3718bbf30cad0ce"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.615820 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.652216 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" event={"ID":"3db9a5a7-2885-4a36-9a69-a38957ddb5a3","Type":"ContainerStarted","Data":"3dabc10c43a615cc0defb5dcb247141f5298d538f340ae46678ee5e58218382c"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.667269 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:52 crc kubenswrapper[4690]: E0217 00:07:52.667973 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:53.167956979 +0000 UTC m=+146.903669810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.684623 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bkbbg" event={"ID":"b65e191f-43ce-4874-82b4-55cebaa7253b","Type":"ContainerStarted","Data":"82bbb63f15d581456d4076080ee69627ba3c42f198c27d90a87c47ba91ba1a41"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.686035 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2nz2\" (UniqueName: \"kubernetes.io/projected/c08acd43-80f4-4a48-b848-7899561e26b5-kube-api-access-b2nz2\") pod \"certified-operators-2tkvq\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.687016 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zjwj9" podStartSLOduration=125.687005175 podStartE2EDuration="2m5.687005175s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:52.651654144 +0000 UTC m=+146.387366905" watchObservedRunningTime="2026-02-17 00:07:52.687005175 +0000 UTC m=+146.422717926" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.687215 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k62p2"] Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.698287 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" event={"ID":"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c","Type":"ContainerStarted","Data":"9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.699644 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.731720 4690 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gxcpk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.731784 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" podUID="4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.736748 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.755154 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" podStartSLOduration=124.755134641 podStartE2EDuration="2m4.755134641s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:52.732175649 +0000 UTC m=+146.467888420" watchObservedRunningTime="2026-02-17 00:07:52.755134641 +0000 UTC m=+146.490847382" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.768861 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" event={"ID":"0ded4d5d-1f0b-4caf-a498-73f7963446fb","Type":"ContainerStarted","Data":"0c52df6c55021087919ca91bba74947d19c171b6df65d791c410a6d2106b6b85"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.769692 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:52 crc kubenswrapper[4690]: E0217 00:07:52.769741 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:53.269726965 +0000 UTC m=+147.005439716 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.792772 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" event={"ID":"56c57ff0-baf4-483e-8763-a3ba52c0130e","Type":"ContainerStarted","Data":"2b89a3925bada73829887de3af72648c2a8efba64d2332b7ac1dc7b54a07fac5"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.793557 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.794557 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pbdwr"] Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.797416 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.811579 4690 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2tt5n container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.811625 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" podUID="56c57ff0-baf4-483e-8763-a3ba52c0130e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.812492 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pbdwr"] Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.812897 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" event={"ID":"73b144f3-32f0-47ca-96dc-54eab0196a50","Type":"ContainerStarted","Data":"373b7ca52ce4859d0260e46e23badea7bc4a58d0f80804f2297732cb3d307704"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.814551 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29521440-6zdhx" event={"ID":"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6","Type":"ContainerStarted","Data":"726ae61d6fb7666db0782798596461846545fe60a1460e9b274ea66f72c94d0d"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.818448 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wgtxp" event={"ID":"d6fdb767-e022-4bdd-8464-753769e47a2a","Type":"ContainerStarted","Data":"c8233d4a6e3d5bc16f69025269ea40e33caf64b13e5f762a46a34aa115921530"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.821659 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vdg7\" (UniqueName: \"kubernetes.io/projected/abbde255-fc8e-4574-b9b2-77e540002308-kube-api-access-8vdg7\") pod \"community-operators-k62p2\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.821722 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.822265 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-catalog-content\") pod \"community-operators-k62p2\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.822316 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-utilities\") pod \"community-operators-k62p2\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:52 crc kubenswrapper[4690]: E0217 00:07:52.824208 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:53.324193375 +0000 UTC m=+147.059906126 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.856184 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" event={"ID":"7c9dd9bc-8d10-4946-a664-d7e5d754c88d","Type":"ContainerStarted","Data":"f56e4b154ef9ca90cddf79f88f2914014891097815663c1ca1a6f035cefd5044"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.856219 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" event={"ID":"7c9dd9bc-8d10-4946-a664-d7e5d754c88d","Type":"ContainerStarted","Data":"569f90688d724c47b963cf7be9b6529bc9b0b603a08906ed12f17054e77f3113"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.887329 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" event={"ID":"0d6d4f41-49a8-4ddd-acfb-1695db85409c","Type":"ContainerStarted","Data":"bf207203e0c4c6141dce63dc9d237a63dd4720f6e742ff1fe3491b231d2cefda"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.887616 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" podStartSLOduration=124.887598701 podStartE2EDuration="2m4.887598701s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:52.845839709 +0000 UTC m=+146.581552460" watchObservedRunningTime="2026-02-17 00:07:52.887598701 +0000 UTC m=+146.623311452" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.917621 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pbbhc" event={"ID":"f06f4ec3-86a3-429b-8920-03dd8e7a4569","Type":"ContainerStarted","Data":"079e531145a4710a9e981a26a956b94b42c075ee0d47d5169313c03d8e201e3f"} Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.919283 4690 patch_prober.go:28] interesting pod/console-operator-58897d9998-pbbhc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.919310 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-pbbhc" podUID="f06f4ec3-86a3-429b-8920-03dd8e7a4569" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.919427 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29521440-6zdhx" podStartSLOduration=125.919411567 podStartE2EDuration="2m5.919411567s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:52.919246202 +0000 UTC m=+146.654958953" watchObservedRunningTime="2026-02-17 00:07:52.919411567 +0000 UTC m=+146.655124308" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.922946 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.923117 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-catalog-content\") pod \"community-operators-k62p2\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.923148 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-catalog-content\") pod \"certified-operators-pbdwr\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.923165 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-utilities\") pod \"community-operators-k62p2\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.923201 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-utilities\") pod \"certified-operators-pbdwr\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.923250 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq88l\" (UniqueName: \"kubernetes.io/projected/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-kube-api-access-cq88l\") pod \"certified-operators-pbdwr\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.923329 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vdg7\" (UniqueName: \"kubernetes.io/projected/abbde255-fc8e-4574-b9b2-77e540002308-kube-api-access-8vdg7\") pod \"community-operators-k62p2\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.924505 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-utilities\") pod \"community-operators-k62p2\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:52 crc kubenswrapper[4690]: E0217 00:07:52.927036 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:53.427016943 +0000 UTC m=+147.162729694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.930637 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-catalog-content\") pod \"community-operators-k62p2\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.950336 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" podStartSLOduration=125.950317886 podStartE2EDuration="2m5.950317886s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:52.946722449 +0000 UTC m=+146.682435200" watchObservedRunningTime="2026-02-17 00:07:52.950317886 +0000 UTC m=+146.686030627" Feb 17 00:07:52 crc kubenswrapper[4690]: I0217 00:07:52.985065 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-687jq" event={"ID":"8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7","Type":"ContainerStarted","Data":"8dcd5f0bbc3e982dfd5e115e09977f772bb92871a5d3b4b71045c80861a6aba8"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.004149 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" podStartSLOduration=125.004126866 podStartE2EDuration="2m5.004126866s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:52.987828001 +0000 UTC m=+146.723540752" watchObservedRunningTime="2026-02-17 00:07:53.004126866 +0000 UTC m=+146.739839617" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.009325 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vdg7\" (UniqueName: \"kubernetes.io/projected/abbde255-fc8e-4574-b9b2-77e540002308-kube-api-access-8vdg7\") pod \"community-operators-k62p2\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.025009 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-catalog-content\") pod \"certified-operators-pbdwr\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.025058 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-utilities\") pod \"certified-operators-pbdwr\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.025093 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq88l\" (UniqueName: \"kubernetes.io/projected/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-kube-api-access-cq88l\") pod \"certified-operators-pbdwr\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.025149 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:53 crc kubenswrapper[4690]: E0217 00:07:53.027139 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:53.52712686 +0000 UTC m=+147.262839611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.027806 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-catalog-content\") pod \"certified-operators-pbdwr\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.038694 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-utilities\") pod \"certified-operators-pbdwr\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.052778 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4h7ff" event={"ID":"e4268b59-7563-4783-9957-2d1656b4ddec","Type":"ContainerStarted","Data":"d3c4014b2625bf408ad606b09588b3dbd197b3a4404384b10bfdc439da1e5af6"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.056668 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.061214 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq88l\" (UniqueName: \"kubernetes.io/projected/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-kube-api-access-cq88l\") pod \"certified-operators-pbdwr\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.074773 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8f7ps" event={"ID":"21ad60ce-9b8d-4dfc-bc60-06901e688248","Type":"ContainerStarted","Data":"a77c5b639e68ef5a2a80002080ea3117c13496b7d3486b7d7cfb8bb9fea261a7"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.109990 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-4h7ff" podStartSLOduration=7.109977614 podStartE2EDuration="7.109977614s" podCreationTimestamp="2026-02-17 00:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:53.107902032 +0000 UTC m=+146.843614783" watchObservedRunningTime="2026-02-17 00:07:53.109977614 +0000 UTC m=+146.845690365" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.110707 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-687jq" podStartSLOduration=126.110703106 podStartE2EDuration="2m6.110703106s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:53.057490503 +0000 UTC m=+146.793203254" watchObservedRunningTime="2026-02-17 00:07:53.110703106 +0000 UTC m=+146.846415857" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.119578 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:07:53 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:07:53 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:07:53 crc kubenswrapper[4690]: healthz check failed Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.119627 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.133482 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" event={"ID":"51b6988f-45a7-4faa-9f25-0c03e4fe316b","Type":"ContainerStarted","Data":"22ca5a526c50ab709c0bfa6a192c03cc1efbdebb979a349a67f1cf9e4a33acb9"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.141072 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:53 crc kubenswrapper[4690]: E0217 00:07:53.141738 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:53.641723388 +0000 UTC m=+147.377436139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.155792 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-17 00:02:52 +0000 UTC, rotation deadline is 2026-12-18 19:02:46.694820658 +0000 UTC Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.155824 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7314h54m53.538999221s for next certificate rotation Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.171679 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.197672 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" event={"ID":"b6070dad-196d-4179-858c-4f75ddb71ebc","Type":"ContainerStarted","Data":"a79aec2120516fe5defbd1293e30c54d2a78f2d38d9fd6fc7133c8be67a978ae"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.197724 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" event={"ID":"b6070dad-196d-4179-858c-4f75ddb71ebc","Type":"ContainerStarted","Data":"d00966aee5802f019921fecc32fc1f89f0135886d5e05373bfaafab785c75f64"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.222502 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" event={"ID":"5604a4e2-1b7b-40a6-9c33-38469fb706b4","Type":"ContainerStarted","Data":"d6c9a4b6a44ccadcb9dc8914a527a0a8aa63b58c2ec9adc40ce02abb40387015"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.237221 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7rclh" podStartSLOduration=125.237198718 podStartE2EDuration="2m5.237198718s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:53.208759432 +0000 UTC m=+146.944472183" watchObservedRunningTime="2026-02-17 00:07:53.237198718 +0000 UTC m=+146.972911469" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.266853 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:53 crc kubenswrapper[4690]: E0217 00:07:53.272380 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:53.772345963 +0000 UTC m=+147.508058714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.274550 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" event={"ID":"983bb1a5-259a-4d98-8c43-4d91d56825bb","Type":"ContainerStarted","Data":"6fda257d2d2d805e9f42f52a0ef7fce6844de3a290ec730744f351df7286b085"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.274622 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" event={"ID":"983bb1a5-259a-4d98-8c43-4d91d56825bb","Type":"ContainerStarted","Data":"4d66afaec79a4a6d9f70741113075d8f1dd6728d390f44f2104c4468e6c9a2d2"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.274936 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.292284 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-66x5j" podStartSLOduration=126.292261765 podStartE2EDuration="2m6.292261765s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:53.264712836 +0000 UTC m=+147.000425597" watchObservedRunningTime="2026-02-17 00:07:53.292261765 +0000 UTC m=+147.027974516" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.327907 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" event={"ID":"ac4bf154-96bf-4052-b8e6-371ede515470","Type":"ContainerStarted","Data":"221061e8e0ab51ee46651ea816ecc754eb1d72e2d456b381bb3cc00c965037df"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.369625 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:53 crc kubenswrapper[4690]: E0217 00:07:53.371422 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:53.871403099 +0000 UTC m=+147.607115850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.413156 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-g8jrf" podStartSLOduration=126.41313733 podStartE2EDuration="2m6.41313733s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:53.411904753 +0000 UTC m=+147.147617504" watchObservedRunningTime="2026-02-17 00:07:53.41313733 +0000 UTC m=+147.148850081" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.436568 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fcmdk"] Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.437067 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" event={"ID":"b0373e24-779c-47ec-b8d7-d4152b995948","Type":"ContainerStarted","Data":"392d973b4f75806278ec89b8d1eda61f30615b29d63f33033cd44167dcfeab1e"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.437110 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" event={"ID":"b0373e24-779c-47ec-b8d7-d4152b995948","Type":"ContainerStarted","Data":"2c6f4bf049f69f8b5935711111322c3a2d28e2bb903de990d54ab4685b72291d"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.460594 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nxjhm" event={"ID":"ab15103a-6de2-47f7-8640-6c3253d2ee68","Type":"ContainerStarted","Data":"e98f9a3522e66249ae900624f6075960fb54ca8ff5136daab9dbb52602ba439c"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.481444 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:53 crc kubenswrapper[4690]: E0217 00:07:53.482505 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:53.982486222 +0000 UTC m=+147.718198973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.495695 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xkbpk" podStartSLOduration=126.495674084 podStartE2EDuration="2m6.495674084s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:53.480777431 +0000 UTC m=+147.216490182" watchObservedRunningTime="2026-02-17 00:07:53.495674084 +0000 UTC m=+147.231386835" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.505586 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nstcz" event={"ID":"a94ddfaa-d52b-4ded-ad8b-c0732e08a983","Type":"ContainerStarted","Data":"6da4d2f8eb1d096b51d5081ca39505d27e56f94d4c4ce0742f403a30765593d3"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.546669 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" event={"ID":"42491e41-75ad-4087-90ea-69ec8d173716","Type":"ContainerStarted","Data":"47a34be16e0a7503d4b68af624c5206a786543e898fedb009722b06c42bb8933"} Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.547384 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" podStartSLOduration=125.547375482 podStartE2EDuration="2m5.547375482s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:53.546931609 +0000 UTC m=+147.282644360" watchObservedRunningTime="2026-02-17 00:07:53.547375482 +0000 UTC m=+147.283088223" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.586352 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.586848 4690 patch_prober.go:28] interesting pod/downloads-7954f5f757-xflvn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.586884 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xflvn" podUID="929321b2-58ac-4633-af87-da506a21c23a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.587800 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" event={"ID":"acc970de-869b-4a13-8d0e-94370922b22f","Type":"ContainerStarted","Data":"3d48a78fe0465eff12f35c7c859b2dda8f205e76e3e1037b3e7fb8e6500871a2"} Feb 17 00:07:53 crc kubenswrapper[4690]: E0217 00:07:53.588927 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:54.088911437 +0000 UTC m=+147.824624188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.614533 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.630016 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" podStartSLOduration=126.629994469 podStartE2EDuration="2m6.629994469s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:53.586748043 +0000 UTC m=+147.322460794" watchObservedRunningTime="2026-02-17 00:07:53.629994469 +0000 UTC m=+147.365707220" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.681734 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6pwz" podStartSLOduration=125.681718887 podStartE2EDuration="2m5.681718887s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:53.628947768 +0000 UTC m=+147.364660519" watchObservedRunningTime="2026-02-17 00:07:53.681718887 +0000 UTC m=+147.417431638" Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.689584 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:53 crc kubenswrapper[4690]: E0217 00:07:53.694390 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:54.194353613 +0000 UTC m=+147.930066364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.792816 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:53 crc kubenswrapper[4690]: E0217 00:07:53.793504 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:54.293489661 +0000 UTC m=+148.029202412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.894453 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:53 crc kubenswrapper[4690]: E0217 00:07:53.894802 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:54.394790424 +0000 UTC m=+148.130503175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.958584 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2tkvq"] Feb 17 00:07:53 crc kubenswrapper[4690]: W0217 00:07:53.987675 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc08acd43_80f4_4a48_b848_7899561e26b5.slice/crio-e15ae248a29acec1068184563d7625566e9ccfb2ee038754f31a628577d5a9ca WatchSource:0}: Error finding container e15ae248a29acec1068184563d7625566e9ccfb2ee038754f31a628577d5a9ca: Status 404 returned error can't find the container with id e15ae248a29acec1068184563d7625566e9ccfb2ee038754f31a628577d5a9ca Feb 17 00:07:53 crc kubenswrapper[4690]: I0217 00:07:53.995194 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:53 crc kubenswrapper[4690]: E0217 00:07:53.995596 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:54.495581171 +0000 UTC m=+148.231293922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.044337 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2j49q" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.047562 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k62p2"] Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.100207 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:54 crc kubenswrapper[4690]: E0217 00:07:54.100837 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:54.600825841 +0000 UTC m=+148.336538592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.104533 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:07:54 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:07:54 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:07:54 crc kubenswrapper[4690]: healthz check failed Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.104574 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.197470 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pv76l"] Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.198422 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.203289 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.203932 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:54 crc kubenswrapper[4690]: E0217 00:07:54.204279 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:54.704266268 +0000 UTC m=+148.439979019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.224150 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pbdwr"] Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.230802 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pv76l"] Feb 17 00:07:54 crc kubenswrapper[4690]: W0217 00:07:54.248293 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda24db2e5_dbd4_40be_bf3b_4ec15f5329bd.slice/crio-56ff16172e3dcc5abc967165749332600dd25fb8657186e8fc2d7b77695f6e82 WatchSource:0}: Error finding container 56ff16172e3dcc5abc967165749332600dd25fb8657186e8fc2d7b77695f6e82: Status 404 returned error can't find the container with id 56ff16172e3dcc5abc967165749332600dd25fb8657186e8fc2d7b77695f6e82 Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.306233 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-catalog-content\") pod \"redhat-marketplace-pv76l\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.306794 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-utilities\") pod \"redhat-marketplace-pv76l\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.306914 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbkvf\" (UniqueName: \"kubernetes.io/projected/55ca888a-bbfa-477d-9769-8eab8a6aee98-kube-api-access-tbkvf\") pod \"redhat-marketplace-pv76l\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.307001 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:54 crc kubenswrapper[4690]: E0217 00:07:54.307368 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:54.807340983 +0000 UTC m=+148.543053734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.409854 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.410036 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-utilities\") pod \"redhat-marketplace-pv76l\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.410092 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbkvf\" (UniqueName: \"kubernetes.io/projected/55ca888a-bbfa-477d-9769-8eab8a6aee98-kube-api-access-tbkvf\") pod \"redhat-marketplace-pv76l\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.410162 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-catalog-content\") pod \"redhat-marketplace-pv76l\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.410543 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-catalog-content\") pod \"redhat-marketplace-pv76l\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.410569 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-utilities\") pod \"redhat-marketplace-pv76l\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: E0217 00:07:54.410607 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:54.910593453 +0000 UTC m=+148.646306204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.482608 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbkvf\" (UniqueName: \"kubernetes.io/projected/55ca888a-bbfa-477d-9769-8eab8a6aee98-kube-api-access-tbkvf\") pod \"redhat-marketplace-pv76l\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.512687 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:54 crc kubenswrapper[4690]: E0217 00:07:54.513026 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.013014259 +0000 UTC m=+148.748727010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.564660 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.580309 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-grqqc"] Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.581684 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.599197 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-grqqc"] Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.605062 4690 generic.go:334] "Generic (PLEG): container finished" podID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerID="b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07" exitCode=0 Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.605755 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcmdk" event={"ID":"4c5d01e4-052c-4541-a385-79a8066bba8f","Type":"ContainerDied","Data":"b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.605780 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcmdk" event={"ID":"4c5d01e4-052c-4541-a385-79a8066bba8f","Type":"ContainerStarted","Data":"05e64fbfb40bd1453488500198377bf85b8081227c06b90443ac70705f7e4fc9"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.612863 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.613616 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:54 crc kubenswrapper[4690]: E0217 00:07:54.614017 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.114003472 +0000 UTC m=+148.849716223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.620716 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" event={"ID":"dd75aefb-75fa-42bf-a3b1-5e0c9fa957f6","Type":"ContainerStarted","Data":"23a6e7317b317c20a5dfc0fc5113b45c0331e4b047f622ab1bddd2b2d0408a6d"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.646343 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" event={"ID":"42491e41-75ad-4087-90ea-69ec8d173716","Type":"ContainerStarted","Data":"affa9afd2880fa7f608b9c16d48ebb7d325c6ff5a038eebe26aa95dae4c3a308"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.646403 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" event={"ID":"42491e41-75ad-4087-90ea-69ec8d173716","Type":"ContainerStarted","Data":"f96cf13c3260c4b96cf8177548c4be85ac686e9fa65188a4d40b4312cefd697d"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.652703 4690 generic.go:334] "Generic (PLEG): container finished" podID="3db9a5a7-2885-4a36-9a69-a38957ddb5a3" containerID="4dd3ecbe1f52f02ef1f45e14323232190396589f1ee6c0a83479f65db04c07eb" exitCode=0 Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.652823 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" event={"ID":"3db9a5a7-2885-4a36-9a69-a38957ddb5a3","Type":"ContainerDied","Data":"4dd3ecbe1f52f02ef1f45e14323232190396589f1ee6c0a83479f65db04c07eb"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.685114 4690 generic.go:334] "Generic (PLEG): container finished" podID="c08acd43-80f4-4a48-b848-7899561e26b5" containerID="97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d" exitCode=0 Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.685539 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2tkvq" event={"ID":"c08acd43-80f4-4a48-b848-7899561e26b5","Type":"ContainerDied","Data":"97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.685624 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2tkvq" event={"ID":"c08acd43-80f4-4a48-b848-7899561e26b5","Type":"ContainerStarted","Data":"e15ae248a29acec1068184563d7625566e9ccfb2ee038754f31a628577d5a9ca"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.696611 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j992c" podStartSLOduration=127.696591728 podStartE2EDuration="2m7.696591728s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:54.689434785 +0000 UTC m=+148.425147536" watchObservedRunningTime="2026-02-17 00:07:54.696591728 +0000 UTC m=+148.432304479" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.702739 4690 generic.go:334] "Generic (PLEG): container finished" podID="abbde255-fc8e-4574-b9b2-77e540002308" containerID="de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37" exitCode=0 Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.702809 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k62p2" event={"ID":"abbde255-fc8e-4574-b9b2-77e540002308","Type":"ContainerDied","Data":"de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.702833 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k62p2" event={"ID":"abbde255-fc8e-4574-b9b2-77e540002308","Type":"ContainerStarted","Data":"f4c2b952dba4dacd3437574ca8b7734676f39447ce23a06f6ace7877d4a0fe37"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.709609 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbdwr" event={"ID":"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd","Type":"ContainerStarted","Data":"ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.709650 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbdwr" event={"ID":"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd","Type":"ContainerStarted","Data":"56ff16172e3dcc5abc967165749332600dd25fb8657186e8fc2d7b77695f6e82"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.714674 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t57sb\" (UniqueName: \"kubernetes.io/projected/573d80ca-54d0-4c1a-8987-e734917e9900-kube-api-access-t57sb\") pod \"redhat-marketplace-grqqc\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.714764 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-utilities\") pod \"redhat-marketplace-grqqc\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.714808 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.714892 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-catalog-content\") pod \"redhat-marketplace-grqqc\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:54 crc kubenswrapper[4690]: E0217 00:07:54.715952 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.215939943 +0000 UTC m=+148.951652694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.717184 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" event={"ID":"0ded4d5d-1f0b-4caf-a498-73f7963446fb","Type":"ContainerStarted","Data":"61ec1c22ed0431b45d9e626efe734ca52843055f44dce36afd4f655890bbd7b2"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.719880 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n8hnq" event={"ID":"7c9dd9bc-8d10-4946-a664-d7e5d754c88d","Type":"ContainerStarted","Data":"420b5a6ff3009a8c396243303676593471e1ea64df8cacdb17e5397eca2953fb"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.725687 4690 generic.go:334] "Generic (PLEG): container finished" podID="0d6d4f41-49a8-4ddd-acfb-1695db85409c" containerID="bf207203e0c4c6141dce63dc9d237a63dd4720f6e742ff1fe3491b231d2cefda" exitCode=0 Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.725760 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" event={"ID":"0d6d4f41-49a8-4ddd-acfb-1695db85409c","Type":"ContainerDied","Data":"bf207203e0c4c6141dce63dc9d237a63dd4720f6e742ff1fe3491b231d2cefda"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.741448 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4pvmr" event={"ID":"73b144f3-32f0-47ca-96dc-54eab0196a50","Type":"ContainerStarted","Data":"d41d18e0cc1c8d8f269f374718b11a21416c85868165a7cc483a8b738aaa7901"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.762523 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2b72" event={"ID":"b0373e24-779c-47ec-b8d7-d4152b995948","Type":"ContainerStarted","Data":"496f01ed2956e2d34b513865e0539dd97f79aa1d715f0ec0eb9103ce2a13c21c"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.776535 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-2jr5f" podStartSLOduration=126.776518835 podStartE2EDuration="2m6.776518835s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:54.775730682 +0000 UTC m=+148.511443433" watchObservedRunningTime="2026-02-17 00:07:54.776518835 +0000 UTC m=+148.512231586" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.819339 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.819480 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-catalog-content\") pod \"redhat-marketplace-grqqc\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.819572 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t57sb\" (UniqueName: \"kubernetes.io/projected/573d80ca-54d0-4c1a-8987-e734917e9900-kube-api-access-t57sb\") pod \"redhat-marketplace-grqqc\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.819704 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-utilities\") pod \"redhat-marketplace-grqqc\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:54 crc kubenswrapper[4690]: E0217 00:07:54.820681 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.320660588 +0000 UTC m=+149.056373339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.821013 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-catalog-content\") pod \"redhat-marketplace-grqqc\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.825520 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-utilities\") pod \"redhat-marketplace-grqqc\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.838266 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wgtxp" event={"ID":"d6fdb767-e022-4bdd-8464-753769e47a2a","Type":"ContainerStarted","Data":"b75dc35a93cc34768ffe0c1468bd47d1574278e587ac0d48cc5e48ce8233c5a1"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.838316 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wgtxp" event={"ID":"d6fdb767-e022-4bdd-8464-753769e47a2a","Type":"ContainerStarted","Data":"710b0a96b48dffbbe681c8f2e6fe864eb4fefa2a4c4d842729068eaecbf00c1b"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.838932 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-wgtxp" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.847119 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vwd2q" podStartSLOduration=127.847103794 podStartE2EDuration="2m7.847103794s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:54.841879049 +0000 UTC m=+148.577591800" watchObservedRunningTime="2026-02-17 00:07:54.847103794 +0000 UTC m=+148.582816545" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.872960 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j28sv" event={"ID":"51a85cef-6a5b-4bde-b9fc-0685d2a96631","Type":"ContainerStarted","Data":"00889651ff88dfd9049b32d2debd07d77a99279ac3cc5ee0bbb7bad58354f928"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.877413 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t57sb\" (UniqueName: \"kubernetes.io/projected/573d80ca-54d0-4c1a-8987-e734917e9900-kube-api-access-t57sb\") pod \"redhat-marketplace-grqqc\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.910183 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" event={"ID":"cd0a26f5-e912-40a5-a41f-3754297b6f68","Type":"ContainerStarted","Data":"1ffc50266e01c4d8f1b98522769537682b6176954f5b9b042eebe8840d16cb1e"} Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.911593 4690 patch_prober.go:28] interesting pod/downloads-7954f5f757-xflvn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.911629 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xflvn" podUID="929321b2-58ac-4633-af87-da506a21c23a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.912247 4690 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gxcpk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.912291 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" podUID="4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Feb 17 00:07:54 crc kubenswrapper[4690]: I0217 00:07:54.921247 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:54 crc kubenswrapper[4690]: E0217 00:07:54.921732 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.421714543 +0000 UTC m=+149.157427294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.022881 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.023071 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.523042166 +0000 UTC m=+149.258754917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.024103 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.024246 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" podStartSLOduration=127.024211871 podStartE2EDuration="2m7.024211871s" podCreationTimestamp="2026-02-17 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:54.997175897 +0000 UTC m=+148.732888648" watchObservedRunningTime="2026-02-17 00:07:55.024211871 +0000 UTC m=+148.759924622" Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.030030 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.530014334 +0000 UTC m=+149.265727085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.061416 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-wgtxp" podStartSLOduration=9.061393727 podStartE2EDuration="9.061393727s" podCreationTimestamp="2026-02-17 00:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:54.957192458 +0000 UTC m=+148.692905209" watchObservedRunningTime="2026-02-17 00:07:55.061393727 +0000 UTC m=+148.797106478" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.106742 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.118346 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:07:55 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:07:55 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:07:55 crc kubenswrapper[4690]: healthz check failed Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.118413 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.127842 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.128306 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.128398 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.128525 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.130209 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.630189413 +0000 UTC m=+149.365902164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.130567 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.139240 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.146201 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.197998 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.209560 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pv76l"] Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.228134 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.232386 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.232511 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.233643 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.733627469 +0000 UTC m=+149.469340220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: W0217 00:07:55.239124 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55ca888a_bbfa_477d_9769_8eab8a6aee98.slice/crio-8a20183b0f15bee1b9ae76d010ce117a6e0a627fb22f79bdd64e81bdf3dd72a0 WatchSource:0}: Error finding container 8a20183b0f15bee1b9ae76d010ce117a6e0a627fb22f79bdd64e81bdf3dd72a0: Status 404 returned error can't find the container with id 8a20183b0f15bee1b9ae76d010ce117a6e0a627fb22f79bdd64e81bdf3dd72a0 Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.240376 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.333060 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.333348 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.833314664 +0000 UTC m=+149.569027415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.435230 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.435704 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:55.935687528 +0000 UTC m=+149.671400279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.462735 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-grqqc"] Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.517060 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.537729 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.537782 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2tt5n" Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.538136 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:56.038121835 +0000 UTC m=+149.773834586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.567681 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xntcs"] Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.568692 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.578863 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.625566 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xntcs"] Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.659393 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.659439 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5v2z\" (UniqueName: \"kubernetes.io/projected/e5df6322-f224-4e0e-8099-02865f10660a-kube-api-access-m5v2z\") pod \"redhat-operators-xntcs\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.659483 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-catalog-content\") pod \"redhat-operators-xntcs\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.659515 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-utilities\") pod \"redhat-operators-xntcs\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.659779 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:56.159768022 +0000 UTC m=+149.895480773 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.761095 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.761329 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5v2z\" (UniqueName: \"kubernetes.io/projected/e5df6322-f224-4e0e-8099-02865f10660a-kube-api-access-m5v2z\") pod \"redhat-operators-xntcs\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.761398 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-catalog-content\") pod \"redhat-operators-xntcs\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.761430 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-utilities\") pod \"redhat-operators-xntcs\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.761538 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:56.261508248 +0000 UTC m=+149.997220999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.762271 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-catalog-content\") pod \"redhat-operators-xntcs\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.762301 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-utilities\") pod \"redhat-operators-xntcs\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.775353 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-pbbhc" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.786107 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5v2z\" (UniqueName: \"kubernetes.io/projected/e5df6322-f224-4e0e-8099-02865f10660a-kube-api-access-m5v2z\") pod \"redhat-operators-xntcs\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.871015 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.871415 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:56.371403846 +0000 UTC m=+150.107116597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.956622 4690 generic.go:334] "Generic (PLEG): container finished" podID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerID="ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae" exitCode=0 Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.956705 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbdwr" event={"ID":"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd","Type":"ContainerDied","Data":"ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae"} Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.972305 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:55 crc kubenswrapper[4690]: E0217 00:07:55.972601 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 00:07:56.472585935 +0000 UTC m=+150.208298686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.976417 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zd95f"] Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.977373 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b2567583033a1b69d52d6ce9bc5fc4ba865e21f4cb337b97dbd76f29f59f4ed3"} Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.977472 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:55 crc kubenswrapper[4690]: I0217 00:07:55.998521 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.007725 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zd95f"] Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.024275 4690 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.030617 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" event={"ID":"3db9a5a7-2885-4a36-9a69-a38957ddb5a3","Type":"ContainerStarted","Data":"8902791acb1b378e65aff7de9d732e59c76f0ceac2186052033dc928e3c1955d"} Feb 17 00:07:56 crc kubenswrapper[4690]: W0217 00:07:56.056471 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-535f8c0757f8cc2cfb981997b3c09c87941e877e6f1cc97ca0c0225528902529 WatchSource:0}: Error finding container 535f8c0757f8cc2cfb981997b3c09c87941e877e6f1cc97ca0c0225528902529: Status 404 returned error can't find the container with id 535f8c0757f8cc2cfb981997b3c09c87941e877e6f1cc97ca0c0225528902529 Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.056735 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j28sv" event={"ID":"51a85cef-6a5b-4bde-b9fc-0685d2a96631","Type":"ContainerStarted","Data":"daf35eed18907b70e4cef26e7a2c9c4173bc481c4928d3363379d8ba12fdf8fd"} Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.058659 4690 generic.go:334] "Generic (PLEG): container finished" podID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerID="75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a" exitCode=0 Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.058725 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pv76l" event={"ID":"55ca888a-bbfa-477d-9769-8eab8a6aee98","Type":"ContainerDied","Data":"75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a"} Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.058750 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pv76l" event={"ID":"55ca888a-bbfa-477d-9769-8eab8a6aee98","Type":"ContainerStarted","Data":"8a20183b0f15bee1b9ae76d010ce117a6e0a627fb22f79bdd64e81bdf3dd72a0"} Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.062151 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grqqc" event={"ID":"573d80ca-54d0-4c1a-8987-e734917e9900","Type":"ContainerStarted","Data":"67607fd6c1d2383b709f039d3cb16d20c24337b8029257f7ebc496017f01cee0"} Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.070793 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.075148 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-catalog-content\") pod \"redhat-operators-zd95f\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.075240 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.075263 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x528p\" (UniqueName: \"kubernetes.io/projected/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-kube-api-access-x528p\") pod \"redhat-operators-zd95f\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.075304 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-utilities\") pod \"redhat-operators-zd95f\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:56 crc kubenswrapper[4690]: E0217 00:07:56.079028 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 00:07:56.579014831 +0000 UTC m=+150.314727582 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9pwzh" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.099468 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:07:56 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:07:56 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:07:56 crc kubenswrapper[4690]: healthz check failed Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.099544 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.138145 4690 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-17T00:07:56.024304104Z","Handler":null,"Name":""} Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.158581 4690 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.158617 4690 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.176479 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.176752 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-utilities\") pod \"redhat-operators-zd95f\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.180847 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-utilities\") pod \"redhat-operators-zd95f\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.181297 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-catalog-content\") pod \"redhat-operators-zd95f\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.181715 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x528p\" (UniqueName: \"kubernetes.io/projected/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-kube-api-access-x528p\") pod \"redhat-operators-zd95f\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.181925 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-catalog-content\") pod \"redhat-operators-zd95f\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.189451 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.217295 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x528p\" (UniqueName: \"kubernetes.io/projected/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-kube-api-access-x528p\") pod \"redhat-operators-zd95f\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:56 crc kubenswrapper[4690]: W0217 00:07:56.282670 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-6c0b486a602d643d22040a3ddb89115855307f7d3a093c61baaa1a68c93ee3de WatchSource:0}: Error finding container 6c0b486a602d643d22040a3ddb89115855307f7d3a093c61baaa1a68c93ee3de: Status 404 returned error can't find the container with id 6c0b486a602d643d22040a3ddb89115855307f7d3a093c61baaa1a68c93ee3de Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.283082 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.329497 4690 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.329541 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.332477 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.332527 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.362564 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.470247 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9pwzh\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.486379 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.508226 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.590811 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d6d4f41-49a8-4ddd-acfb-1695db85409c-config-volume\") pod \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.590854 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d6d4f41-49a8-4ddd-acfb-1695db85409c-secret-volume\") pod \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.590904 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5q78\" (UniqueName: \"kubernetes.io/projected/0d6d4f41-49a8-4ddd-acfb-1695db85409c-kube-api-access-c5q78\") pod \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\" (UID: \"0d6d4f41-49a8-4ddd-acfb-1695db85409c\") " Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.595852 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d6d4f41-49a8-4ddd-acfb-1695db85409c-config-volume" (OuterVolumeSpecName: "config-volume") pod "0d6d4f41-49a8-4ddd-acfb-1695db85409c" (UID: "0d6d4f41-49a8-4ddd-acfb-1695db85409c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.613000 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d6d4f41-49a8-4ddd-acfb-1695db85409c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0d6d4f41-49a8-4ddd-acfb-1695db85409c" (UID: "0d6d4f41-49a8-4ddd-acfb-1695db85409c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.620537 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d6d4f41-49a8-4ddd-acfb-1695db85409c-kube-api-access-c5q78" (OuterVolumeSpecName: "kube-api-access-c5q78") pod "0d6d4f41-49a8-4ddd-acfb-1695db85409c" (UID: "0d6d4f41-49a8-4ddd-acfb-1695db85409c"). InnerVolumeSpecName "kube-api-access-c5q78". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.692335 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d6d4f41-49a8-4ddd-acfb-1695db85409c-config-volume\") on node \"crc\" DevicePath \"\"" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.692687 4690 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d6d4f41-49a8-4ddd-acfb-1695db85409c-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.692697 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5q78\" (UniqueName: \"kubernetes.io/projected/0d6d4f41-49a8-4ddd-acfb-1695db85409c-kube-api-access-c5q78\") on node \"crc\" DevicePath \"\"" Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.792538 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xntcs"] Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.928636 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9pwzh"] Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.951485 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zd95f"] Feb 17 00:07:56 crc kubenswrapper[4690]: I0217 00:07:56.997931 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.093693 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"83fc848cd070aecf0195a999c4d509742cd8105b1b934a33bd5096c12bc23242"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.100682 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:07:57 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:07:57 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:07:57 crc kubenswrapper[4690]: healthz check failed Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.100737 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.107115 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"fecba1a8f3b92229e88e09baa85850a5e0cbde0516daa2b5c53a639feadc07f7"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.107161 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"535f8c0757f8cc2cfb981997b3c09c87941e877e6f1cc97ca0c0225528902529"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.113035 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd95f" event={"ID":"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2","Type":"ContainerStarted","Data":"05e215f7b413970df290bee3e1ade448f1dd264cc6b8a68bc7ebb78582889507"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.150582 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j28sv" event={"ID":"51a85cef-6a5b-4bde-b9fc-0685d2a96631","Type":"ContainerStarted","Data":"d99ef47707f3799428ffb4eb13686d2d941922e4161ced0506287623a0b8ecdb"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.150645 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j28sv" event={"ID":"51a85cef-6a5b-4bde-b9fc-0685d2a96631","Type":"ContainerStarted","Data":"a1804895c4e7269d0b2f9c4f79e6430e6bf875f0dd4a2bb542b7344a26406c5f"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.154085 4690 generic.go:334] "Generic (PLEG): container finished" podID="573d80ca-54d0-4c1a-8987-e734917e9900" containerID="da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092" exitCode=0 Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.154420 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grqqc" event={"ID":"573d80ca-54d0-4c1a-8987-e734917e9900","Type":"ContainerDied","Data":"da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.167414 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" event={"ID":"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e","Type":"ContainerStarted","Data":"55313a5fee0f3cd37f3a73b837c05d53ee3d188b579e8a870c2597370bd4939d"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.187210 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" event={"ID":"3db9a5a7-2885-4a36-9a69-a38957ddb5a3","Type":"ContainerStarted","Data":"e9139462c390338a2a4448c825f4295b045ea38743344cc961c66b1b2109dbb4"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.206248 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-j28sv" podStartSLOduration=11.206228993 podStartE2EDuration="11.206228993s" podCreationTimestamp="2026-02-17 00:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:57.205466211 +0000 UTC m=+150.941178962" watchObservedRunningTime="2026-02-17 00:07:57.206228993 +0000 UTC m=+150.941941744" Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.207217 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xntcs" event={"ID":"e5df6322-f224-4e0e-8099-02865f10660a","Type":"ContainerStarted","Data":"7b7ad9439d0a6c91a7feb314bf32ca6c563429bc3b3d5109067561e6ac6e56c1"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.230530 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" event={"ID":"0d6d4f41-49a8-4ddd-acfb-1695db85409c","Type":"ContainerDied","Data":"8af25c56f41de60a150e3368f9202e974ac80fbbc1fb0a11d7337cdd9b99400c"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.230608 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8af25c56f41de60a150e3368f9202e974ac80fbbc1fb0a11d7337cdd9b99400c" Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.230944 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29521440-8z9bb" Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.235186 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"07f0d7f0f59ffc47b380338f08526604d69a836183c90c616f152908eee71680"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.235258 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6c0b486a602d643d22040a3ddb89115855307f7d3a093c61baaa1a68c93ee3de"} Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.235822 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:07:57 crc kubenswrapper[4690]: I0217 00:07:57.251561 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" podStartSLOduration=130.251542951 podStartE2EDuration="2m10.251542951s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:57.248980665 +0000 UTC m=+150.984693426" watchObservedRunningTime="2026-02-17 00:07:57.251542951 +0000 UTC m=+150.987255702" Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.100397 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:07:58 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:07:58 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:07:58 crc kubenswrapper[4690]: healthz check failed Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.100687 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.254786 4690 generic.go:334] "Generic (PLEG): container finished" podID="e5df6322-f224-4e0e-8099-02865f10660a" containerID="96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306" exitCode=0 Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.254860 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xntcs" event={"ID":"e5df6322-f224-4e0e-8099-02865f10660a","Type":"ContainerDied","Data":"96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306"} Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.260872 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" event={"ID":"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e","Type":"ContainerStarted","Data":"41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b"} Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.261003 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.274475 4690 generic.go:334] "Generic (PLEG): container finished" podID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerID="f0b15c526fdeec1ff5cd1c5f04abd6c4010b4df1e9788184b131bdb37d4d78a5" exitCode=0 Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.276004 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd95f" event={"ID":"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2","Type":"ContainerDied","Data":"f0b15c526fdeec1ff5cd1c5f04abd6c4010b4df1e9788184b131bdb37d4d78a5"} Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.308778 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" podStartSLOduration=131.308761401 podStartE2EDuration="2m11.308761401s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:07:58.306802513 +0000 UTC m=+152.042515284" watchObservedRunningTime="2026-02-17 00:07:58.308761401 +0000 UTC m=+152.044474152" Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.992868 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:58 crc kubenswrapper[4690]: I0217 00:07:58.993478 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.000938 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.027708 4690 patch_prober.go:28] interesting pod/downloads-7954f5f757-xflvn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.027749 4690 patch_prober.go:28] interesting pod/downloads-7954f5f757-xflvn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.027774 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xflvn" podUID="929321b2-58ac-4633-af87-da506a21c23a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.027794 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xflvn" podUID="929321b2-58ac-4633-af87-da506a21c23a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.094318 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.096545 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:07:59 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:07:59 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:07:59 crc kubenswrapper[4690]: healthz check failed Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.096587 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.336963 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8sp7g" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.535694 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.535757 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.543783 4690 patch_prober.go:28] interesting pod/console-f9d7485db-687jq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.543843 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-687jq" podUID="8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.867152 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.867203 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:07:59 crc kubenswrapper[4690]: I0217 00:07:59.878977 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.096294 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:00 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:00 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:00 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.096345 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.285208 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 17 00:08:00 crc kubenswrapper[4690]: E0217 00:08:00.285443 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d6d4f41-49a8-4ddd-acfb-1695db85409c" containerName="collect-profiles" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.285454 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d6d4f41-49a8-4ddd-acfb-1695db85409c" containerName="collect-profiles" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.285541 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d6d4f41-49a8-4ddd-acfb-1695db85409c" containerName="collect-profiles" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.285951 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.287973 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.288187 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.294495 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.330441 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-rdw2m" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.468832 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.469456 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.473740 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.473931 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.483606 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5d7bcdfb-faee-42e5-ba23-3ff702fa0545\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.484571 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5d7bcdfb-faee-42e5-ba23-3ff702fa0545\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.495764 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.585478 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5d7bcdfb-faee-42e5-ba23-3ff702fa0545\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.585549 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f6a35a33-f5bf-4ec9-b954-c025466d3ba6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.585911 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5d7bcdfb-faee-42e5-ba23-3ff702fa0545\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.586017 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5d7bcdfb-faee-42e5-ba23-3ff702fa0545\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.586120 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f6a35a33-f5bf-4ec9-b954-c025466d3ba6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.603115 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5d7bcdfb-faee-42e5-ba23-3ff702fa0545\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.605950 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.695907 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f6a35a33-f5bf-4ec9-b954-c025466d3ba6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.696395 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f6a35a33-f5bf-4ec9-b954-c025466d3ba6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.696729 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f6a35a33-f5bf-4ec9-b954-c025466d3ba6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.713527 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f6a35a33-f5bf-4ec9-b954-c025466d3ba6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 00:08:00 crc kubenswrapper[4690]: I0217 00:08:00.793687 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 00:08:01 crc kubenswrapper[4690]: I0217 00:08:01.098191 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:01 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:01 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:01 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:01 crc kubenswrapper[4690]: I0217 00:08:01.098474 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:01 crc kubenswrapper[4690]: I0217 00:08:01.124000 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 17 00:08:01 crc kubenswrapper[4690]: I0217 00:08:01.147842 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 17 00:08:01 crc kubenswrapper[4690]: W0217 00:08:01.161324 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podf6a35a33_f5bf_4ec9_b954_c025466d3ba6.slice/crio-0bfa429bd45e1c734e47121f4dfea2cabd7b3a58c158d95a6e8c3d837bb35e70 WatchSource:0}: Error finding container 0bfa429bd45e1c734e47121f4dfea2cabd7b3a58c158d95a6e8c3d837bb35e70: Status 404 returned error can't find the container with id 0bfa429bd45e1c734e47121f4dfea2cabd7b3a58c158d95a6e8c3d837bb35e70 Feb 17 00:08:01 crc kubenswrapper[4690]: W0217 00:08:01.193124 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5d7bcdfb_faee_42e5_ba23_3ff702fa0545.slice/crio-0a7d16439a14d8e7eae3fac8815c30bb411ac6a05885dad6521904d66ad4e758 WatchSource:0}: Error finding container 0a7d16439a14d8e7eae3fac8815c30bb411ac6a05885dad6521904d66ad4e758: Status 404 returned error can't find the container with id 0a7d16439a14d8e7eae3fac8815c30bb411ac6a05885dad6521904d66ad4e758 Feb 17 00:08:01 crc kubenswrapper[4690]: I0217 00:08:01.347542 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5d7bcdfb-faee-42e5-ba23-3ff702fa0545","Type":"ContainerStarted","Data":"0a7d16439a14d8e7eae3fac8815c30bb411ac6a05885dad6521904d66ad4e758"} Feb 17 00:08:01 crc kubenswrapper[4690]: I0217 00:08:01.359947 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f6a35a33-f5bf-4ec9-b954-c025466d3ba6","Type":"ContainerStarted","Data":"0bfa429bd45e1c734e47121f4dfea2cabd7b3a58c158d95a6e8c3d837bb35e70"} Feb 17 00:08:02 crc kubenswrapper[4690]: I0217 00:08:02.117413 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:02 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:02 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:02 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:02 crc kubenswrapper[4690]: I0217 00:08:02.117737 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:02 crc kubenswrapper[4690]: I0217 00:08:02.402794 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f6a35a33-f5bf-4ec9-b954-c025466d3ba6","Type":"ContainerStarted","Data":"26ff1efe58ef9f1512bac0cc1cda4f0341958463102752b054ebcec0f787b62e"} Feb 17 00:08:02 crc kubenswrapper[4690]: I0217 00:08:02.414350 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5d7bcdfb-faee-42e5-ba23-3ff702fa0545","Type":"ContainerStarted","Data":"6ee6e162a31cbcec52c434e947b5a4b48c20daa88662fb812edb44d5c16756cf"} Feb 17 00:08:02 crc kubenswrapper[4690]: I0217 00:08:02.436073 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.436055634 podStartE2EDuration="2.436055634s" podCreationTimestamp="2026-02-17 00:08:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:08:02.420762079 +0000 UTC m=+156.156474830" watchObservedRunningTime="2026-02-17 00:08:02.436055634 +0000 UTC m=+156.171768385" Feb 17 00:08:02 crc kubenswrapper[4690]: I0217 00:08:02.437432 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.437426444 podStartE2EDuration="2.437426444s" podCreationTimestamp="2026-02-17 00:08:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:08:02.435515388 +0000 UTC m=+156.171228139" watchObservedRunningTime="2026-02-17 00:08:02.437426444 +0000 UTC m=+156.173139195" Feb 17 00:08:03 crc kubenswrapper[4690]: I0217 00:08:03.095035 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:03 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:03 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:03 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:03 crc kubenswrapper[4690]: I0217 00:08:03.095098 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:03 crc kubenswrapper[4690]: I0217 00:08:03.434875 4690 generic.go:334] "Generic (PLEG): container finished" podID="5d7bcdfb-faee-42e5-ba23-3ff702fa0545" containerID="6ee6e162a31cbcec52c434e947b5a4b48c20daa88662fb812edb44d5c16756cf" exitCode=0 Feb 17 00:08:03 crc kubenswrapper[4690]: I0217 00:08:03.434927 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5d7bcdfb-faee-42e5-ba23-3ff702fa0545","Type":"ContainerDied","Data":"6ee6e162a31cbcec52c434e947b5a4b48c20daa88662fb812edb44d5c16756cf"} Feb 17 00:08:03 crc kubenswrapper[4690]: I0217 00:08:03.439437 4690 generic.go:334] "Generic (PLEG): container finished" podID="f6a35a33-f5bf-4ec9-b954-c025466d3ba6" containerID="26ff1efe58ef9f1512bac0cc1cda4f0341958463102752b054ebcec0f787b62e" exitCode=0 Feb 17 00:08:03 crc kubenswrapper[4690]: I0217 00:08:03.439491 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f6a35a33-f5bf-4ec9-b954-c025466d3ba6","Type":"ContainerDied","Data":"26ff1efe58ef9f1512bac0cc1cda4f0341958463102752b054ebcec0f787b62e"} Feb 17 00:08:04 crc kubenswrapper[4690]: I0217 00:08:04.095770 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:04 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:04 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:04 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:04 crc kubenswrapper[4690]: I0217 00:08:04.096134 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:04 crc kubenswrapper[4690]: I0217 00:08:04.718953 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-wgtxp" Feb 17 00:08:05 crc kubenswrapper[4690]: I0217 00:08:05.097223 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:05 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:05 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:05 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:05 crc kubenswrapper[4690]: I0217 00:08:05.097554 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:06 crc kubenswrapper[4690]: I0217 00:08:06.095144 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:06 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:06 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:06 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:06 crc kubenswrapper[4690]: I0217 00:08:06.095297 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:07 crc kubenswrapper[4690]: I0217 00:08:07.096421 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:07 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:07 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:07 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:07 crc kubenswrapper[4690]: I0217 00:08:07.096671 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:08 crc kubenswrapper[4690]: I0217 00:08:08.096514 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:08 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:08 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:08 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:08 crc kubenswrapper[4690]: I0217 00:08:08.096567 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:09 crc kubenswrapper[4690]: I0217 00:08:09.041682 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-xflvn" Feb 17 00:08:09 crc kubenswrapper[4690]: I0217 00:08:09.096191 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:09 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:09 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:09 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:09 crc kubenswrapper[4690]: I0217 00:08:09.096236 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:09 crc kubenswrapper[4690]: I0217 00:08:09.535854 4690 patch_prober.go:28] interesting pod/console-f9d7485db-687jq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Feb 17 00:08:09 crc kubenswrapper[4690]: I0217 00:08:09.535918 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-687jq" podUID="8f8ed95f-8d4b-4975-b8d0-7d7e0f9a9ad7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Feb 17 00:08:10 crc kubenswrapper[4690]: I0217 00:08:10.095531 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:10 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:10 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:10 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:10 crc kubenswrapper[4690]: I0217 00:08:10.095589 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:11 crc kubenswrapper[4690]: I0217 00:08:11.089956 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:08:11 crc kubenswrapper[4690]: I0217 00:08:11.098262 4690 patch_prober.go:28] interesting pod/router-default-5444994796-glck9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 00:08:11 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Feb 17 00:08:11 crc kubenswrapper[4690]: [+]process-running ok Feb 17 00:08:11 crc kubenswrapper[4690]: healthz check failed Feb 17 00:08:11 crc kubenswrapper[4690]: I0217 00:08:11.098345 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glck9" podUID="f56490e5-1c42-4126-b7f5-cd7ef73b32bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 00:08:11 crc kubenswrapper[4690]: I0217 00:08:11.104156 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9050c3d5-4d74-4b57-afba-1dd177ce7983-metrics-certs\") pod \"network-metrics-daemon-lz646\" (UID: \"9050c3d5-4d74-4b57-afba-1dd177ce7983\") " pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:08:11 crc kubenswrapper[4690]: I0217 00:08:11.295874 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lz646" Feb 17 00:08:12 crc kubenswrapper[4690]: I0217 00:08:12.097817 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:08:12 crc kubenswrapper[4690]: I0217 00:08:12.101194 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-glck9" Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.890599 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.897712 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.946736 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kubelet-dir\") pod \"5d7bcdfb-faee-42e5-ba23-3ff702fa0545\" (UID: \"5d7bcdfb-faee-42e5-ba23-3ff702fa0545\") " Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.946868 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5d7bcdfb-faee-42e5-ba23-3ff702fa0545" (UID: "5d7bcdfb-faee-42e5-ba23-3ff702fa0545"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.946910 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kubelet-dir\") pod \"f6a35a33-f5bf-4ec9-b954-c025466d3ba6\" (UID: \"f6a35a33-f5bf-4ec9-b954-c025466d3ba6\") " Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.946974 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kube-api-access\") pod \"f6a35a33-f5bf-4ec9-b954-c025466d3ba6\" (UID: \"f6a35a33-f5bf-4ec9-b954-c025466d3ba6\") " Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.947000 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f6a35a33-f5bf-4ec9-b954-c025466d3ba6" (UID: "f6a35a33-f5bf-4ec9-b954-c025466d3ba6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.947062 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kube-api-access\") pod \"5d7bcdfb-faee-42e5-ba23-3ff702fa0545\" (UID: \"5d7bcdfb-faee-42e5-ba23-3ff702fa0545\") " Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.947895 4690 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.947948 4690 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.951463 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5d7bcdfb-faee-42e5-ba23-3ff702fa0545" (UID: "5d7bcdfb-faee-42e5-ba23-3ff702fa0545"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:08:14 crc kubenswrapper[4690]: I0217 00:08:14.955176 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f6a35a33-f5bf-4ec9-b954-c025466d3ba6" (UID: "f6a35a33-f5bf-4ec9-b954-c025466d3ba6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:08:15 crc kubenswrapper[4690]: I0217 00:08:15.049076 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6a35a33-f5bf-4ec9-b954-c025466d3ba6-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:15 crc kubenswrapper[4690]: I0217 00:08:15.049107 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d7bcdfb-faee-42e5-ba23-3ff702fa0545-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:15 crc kubenswrapper[4690]: I0217 00:08:15.531433 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f6a35a33-f5bf-4ec9-b954-c025466d3ba6","Type":"ContainerDied","Data":"0bfa429bd45e1c734e47121f4dfea2cabd7b3a58c158d95a6e8c3d837bb35e70"} Feb 17 00:08:15 crc kubenswrapper[4690]: I0217 00:08:15.531491 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bfa429bd45e1c734e47121f4dfea2cabd7b3a58c158d95a6e8c3d837bb35e70" Feb 17 00:08:15 crc kubenswrapper[4690]: I0217 00:08:15.531482 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 00:08:15 crc kubenswrapper[4690]: I0217 00:08:15.534901 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5d7bcdfb-faee-42e5-ba23-3ff702fa0545","Type":"ContainerDied","Data":"0a7d16439a14d8e7eae3fac8815c30bb411ac6a05885dad6521904d66ad4e758"} Feb 17 00:08:15 crc kubenswrapper[4690]: I0217 00:08:15.534982 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a7d16439a14d8e7eae3fac8815c30bb411ac6a05885dad6521904d66ad4e758" Feb 17 00:08:15 crc kubenswrapper[4690]: I0217 00:08:15.534987 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 00:08:16 crc kubenswrapper[4690]: I0217 00:08:16.501400 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:08:19 crc kubenswrapper[4690]: I0217 00:08:19.540662 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:08:19 crc kubenswrapper[4690]: I0217 00:08:19.545646 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-687jq" Feb 17 00:08:25 crc kubenswrapper[4690]: I0217 00:08:25.614437 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6" containerID="726ae61d6fb7666db0782798596461846545fe60a1460e9b274ea66f72c94d0d" exitCode=0 Feb 17 00:08:25 crc kubenswrapper[4690]: I0217 00:08:25.614519 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29521440-6zdhx" event={"ID":"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6","Type":"ContainerDied","Data":"726ae61d6fb7666db0782798596461846545fe60a1460e9b274ea66f72c94d0d"} Feb 17 00:08:25 crc kubenswrapper[4690]: I0217 00:08:25.828491 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-c6pc7"] Feb 17 00:08:26 crc kubenswrapper[4690]: I0217 00:08:26.325348 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:08:26 crc kubenswrapper[4690]: I0217 00:08:26.325438 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.151212 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29521440-6zdhx" Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.269336 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wxst\" (UniqueName: \"kubernetes.io/projected/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-kube-api-access-9wxst\") pod \"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6\" (UID: \"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6\") " Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.269636 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-serviceca\") pod \"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6\" (UID: \"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6\") " Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.271319 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-serviceca" (OuterVolumeSpecName: "serviceca") pod "c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6" (UID: "c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.285026 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-kube-api-access-9wxst" (OuterVolumeSpecName: "kube-api-access-9wxst") pod "c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6" (UID: "c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6"). InnerVolumeSpecName "kube-api-access-9wxst". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.370830 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wxst\" (UniqueName: \"kubernetes.io/projected/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-kube-api-access-9wxst\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.372186 4690 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6-serviceca\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.508205 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lz646"] Feb 17 00:08:28 crc kubenswrapper[4690]: W0217 00:08:28.522252 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9050c3d5_4d74_4b57_afba_1dd177ce7983.slice/crio-70ae3861ce93a81264d7ab2c972a31b24624eeb3bc7afab2897a932be0100f8a WatchSource:0}: Error finding container 70ae3861ce93a81264d7ab2c972a31b24624eeb3bc7afab2897a932be0100f8a: Status 404 returned error can't find the container with id 70ae3861ce93a81264d7ab2c972a31b24624eeb3bc7afab2897a932be0100f8a Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.639958 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29521440-6zdhx" event={"ID":"c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6","Type":"ContainerDied","Data":"c5c229e4667e388858094b7cf9f5cc1ae1358905c1be1fb0a8289738a65fd4b4"} Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.640003 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5c229e4667e388858094b7cf9f5cc1ae1358905c1be1fb0a8289738a65fd4b4" Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.640071 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29521440-6zdhx" Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.646866 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd95f" event={"ID":"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2","Type":"ContainerStarted","Data":"8266250e18fe1fc635cde0f323106e87f3250831d64b4c0886160dfc1f60bae7"} Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.654728 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcmdk" event={"ID":"4c5d01e4-052c-4541-a385-79a8066bba8f","Type":"ContainerStarted","Data":"eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459"} Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.658648 4690 generic.go:334] "Generic (PLEG): container finished" podID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerID="e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf" exitCode=0 Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.658740 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pv76l" event={"ID":"55ca888a-bbfa-477d-9769-8eab8a6aee98","Type":"ContainerDied","Data":"e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf"} Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.668487 4690 generic.go:334] "Generic (PLEG): container finished" podID="573d80ca-54d0-4c1a-8987-e734917e9900" containerID="b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459" exitCode=0 Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.668589 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grqqc" event={"ID":"573d80ca-54d0-4c1a-8987-e734917e9900","Type":"ContainerDied","Data":"b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459"} Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.671709 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k62p2" event={"ID":"abbde255-fc8e-4574-b9b2-77e540002308","Type":"ContainerStarted","Data":"aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9"} Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.673091 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbdwr" event={"ID":"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd","Type":"ContainerStarted","Data":"c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec"} Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.674287 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lz646" event={"ID":"9050c3d5-4d74-4b57-afba-1dd177ce7983","Type":"ContainerStarted","Data":"70ae3861ce93a81264d7ab2c972a31b24624eeb3bc7afab2897a932be0100f8a"} Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.677385 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xntcs" event={"ID":"e5df6322-f224-4e0e-8099-02865f10660a","Type":"ContainerStarted","Data":"e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82"} Feb 17 00:08:28 crc kubenswrapper[4690]: I0217 00:08:28.681618 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2tkvq" event={"ID":"c08acd43-80f4-4a48-b848-7899561e26b5","Type":"ContainerStarted","Data":"5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4"} Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.584853 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wslzr" Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.691924 4690 generic.go:334] "Generic (PLEG): container finished" podID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerID="8266250e18fe1fc635cde0f323106e87f3250831d64b4c0886160dfc1f60bae7" exitCode=0 Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.692160 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd95f" event={"ID":"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2","Type":"ContainerDied","Data":"8266250e18fe1fc635cde0f323106e87f3250831d64b4c0886160dfc1f60bae7"} Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.694536 4690 generic.go:334] "Generic (PLEG): container finished" podID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerID="eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459" exitCode=0 Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.694613 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcmdk" event={"ID":"4c5d01e4-052c-4541-a385-79a8066bba8f","Type":"ContainerDied","Data":"eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459"} Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.698565 4690 generic.go:334] "Generic (PLEG): container finished" podID="e5df6322-f224-4e0e-8099-02865f10660a" containerID="e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82" exitCode=0 Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.698635 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xntcs" event={"ID":"e5df6322-f224-4e0e-8099-02865f10660a","Type":"ContainerDied","Data":"e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82"} Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.700979 4690 generic.go:334] "Generic (PLEG): container finished" podID="c08acd43-80f4-4a48-b848-7899561e26b5" containerID="5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4" exitCode=0 Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.701035 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2tkvq" event={"ID":"c08acd43-80f4-4a48-b848-7899561e26b5","Type":"ContainerDied","Data":"5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4"} Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.712469 4690 generic.go:334] "Generic (PLEG): container finished" podID="abbde255-fc8e-4574-b9b2-77e540002308" containerID="aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9" exitCode=0 Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.712551 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k62p2" event={"ID":"abbde255-fc8e-4574-b9b2-77e540002308","Type":"ContainerDied","Data":"aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9"} Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.715516 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lz646" event={"ID":"9050c3d5-4d74-4b57-afba-1dd177ce7983","Type":"ContainerStarted","Data":"15079e93774a42b116874433eb0c16cf1ddda94590d357e08297e6e399d47cf0"} Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.715550 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lz646" event={"ID":"9050c3d5-4d74-4b57-afba-1dd177ce7983","Type":"ContainerStarted","Data":"13aa3609498c2a88baae4a5ec249422c3d45bb13e7aa3b4c172e642ad316dfad"} Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.722709 4690 generic.go:334] "Generic (PLEG): container finished" podID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerID="c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec" exitCode=0 Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.722758 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbdwr" event={"ID":"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd","Type":"ContainerDied","Data":"c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec"} Feb 17 00:08:29 crc kubenswrapper[4690]: I0217 00:08:29.817669 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-lz646" podStartSLOduration=162.8176459 podStartE2EDuration="2m42.8176459s" podCreationTimestamp="2026-02-17 00:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:08:29.812139126 +0000 UTC m=+183.547851867" watchObservedRunningTime="2026-02-17 00:08:29.8176459 +0000 UTC m=+183.553358661" Feb 17 00:08:30 crc kubenswrapper[4690]: I0217 00:08:30.729772 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pv76l" event={"ID":"55ca888a-bbfa-477d-9769-8eab8a6aee98","Type":"ContainerStarted","Data":"8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a"} Feb 17 00:08:30 crc kubenswrapper[4690]: I0217 00:08:30.746658 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pv76l" podStartSLOduration=2.617162435 podStartE2EDuration="36.746643347s" podCreationTimestamp="2026-02-17 00:07:54 +0000 UTC" firstStartedPulling="2026-02-17 00:07:56.063096127 +0000 UTC m=+149.798808868" lastFinishedPulling="2026-02-17 00:08:30.192577029 +0000 UTC m=+183.928289780" observedRunningTime="2026-02-17 00:08:30.74472567 +0000 UTC m=+184.480438431" watchObservedRunningTime="2026-02-17 00:08:30.746643347 +0000 UTC m=+184.482356098" Feb 17 00:08:31 crc kubenswrapper[4690]: I0217 00:08:31.738261 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grqqc" event={"ID":"573d80ca-54d0-4c1a-8987-e734917e9900","Type":"ContainerStarted","Data":"091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34"} Feb 17 00:08:31 crc kubenswrapper[4690]: I0217 00:08:31.758130 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-grqqc" podStartSLOduration=3.993356083 podStartE2EDuration="37.758110117s" podCreationTimestamp="2026-02-17 00:07:54 +0000 UTC" firstStartedPulling="2026-02-17 00:07:57.16339931 +0000 UTC m=+150.899112061" lastFinishedPulling="2026-02-17 00:08:30.928153344 +0000 UTC m=+184.663866095" observedRunningTime="2026-02-17 00:08:31.757165529 +0000 UTC m=+185.492878270" watchObservedRunningTime="2026-02-17 00:08:31.758110117 +0000 UTC m=+185.493822868" Feb 17 00:08:33 crc kubenswrapper[4690]: I0217 00:08:33.754932 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcmdk" event={"ID":"4c5d01e4-052c-4541-a385-79a8066bba8f","Type":"ContainerStarted","Data":"5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec"} Feb 17 00:08:34 crc kubenswrapper[4690]: I0217 00:08:34.565574 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:08:34 crc kubenswrapper[4690]: I0217 00:08:34.565886 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:08:34 crc kubenswrapper[4690]: I0217 00:08:34.785040 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fcmdk" podStartSLOduration=4.132765239 podStartE2EDuration="42.785024615s" podCreationTimestamp="2026-02-17 00:07:52 +0000 UTC" firstStartedPulling="2026-02-17 00:07:54.612559569 +0000 UTC m=+148.348272320" lastFinishedPulling="2026-02-17 00:08:33.264818945 +0000 UTC m=+187.000531696" observedRunningTime="2026-02-17 00:08:34.782088747 +0000 UTC m=+188.517801508" watchObservedRunningTime="2026-02-17 00:08:34.785024615 +0000 UTC m=+188.520737356" Feb 17 00:08:35 crc kubenswrapper[4690]: I0217 00:08:35.107810 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:08:35 crc kubenswrapper[4690]: I0217 00:08:35.107863 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:08:35 crc kubenswrapper[4690]: I0217 00:08:35.496975 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:08:35 crc kubenswrapper[4690]: I0217 00:08:35.497389 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:08:35 crc kubenswrapper[4690]: I0217 00:08:35.647511 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 00:08:35 crc kubenswrapper[4690]: I0217 00:08:35.766712 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd95f" event={"ID":"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2","Type":"ContainerStarted","Data":"f8feb44d671255344b6c5e4472af6be49efaa0f20aadd8771a9cec6866973bf7"} Feb 17 00:08:35 crc kubenswrapper[4690]: I0217 00:08:35.801339 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.683384 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 17 00:08:36 crc kubenswrapper[4690]: E0217 00:08:36.683696 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6" containerName="image-pruner" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.683712 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6" containerName="image-pruner" Feb 17 00:08:36 crc kubenswrapper[4690]: E0217 00:08:36.683727 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6a35a33-f5bf-4ec9-b954-c025466d3ba6" containerName="pruner" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.683736 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6a35a33-f5bf-4ec9-b954-c025466d3ba6" containerName="pruner" Feb 17 00:08:36 crc kubenswrapper[4690]: E0217 00:08:36.683749 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d7bcdfb-faee-42e5-ba23-3ff702fa0545" containerName="pruner" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.683759 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d7bcdfb-faee-42e5-ba23-3ff702fa0545" containerName="pruner" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.683885 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8a090bc-96a0-4c1c-9cc3-1634bbcfa3e6" containerName="image-pruner" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.683905 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d7bcdfb-faee-42e5-ba23-3ff702fa0545" containerName="pruner" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.683916 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6a35a33-f5bf-4ec9-b954-c025466d3ba6" containerName="pruner" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.684389 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.686864 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.686864 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.695921 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.775580 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2tkvq" event={"ID":"c08acd43-80f4-4a48-b848-7899561e26b5","Type":"ContainerStarted","Data":"7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2"} Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.784883 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7423f424-b2b9-4045-8dce-9417d3ad6eab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7423f424-b2b9-4045-8dce-9417d3ad6eab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.787144 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7423f424-b2b9-4045-8dce-9417d3ad6eab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7423f424-b2b9-4045-8dce-9417d3ad6eab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.802227 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zd95f" podStartSLOduration=5.180817765 podStartE2EDuration="41.802199544s" podCreationTimestamp="2026-02-17 00:07:55 +0000 UTC" firstStartedPulling="2026-02-17 00:07:58.277606825 +0000 UTC m=+152.013319576" lastFinishedPulling="2026-02-17 00:08:34.898988604 +0000 UTC m=+188.634701355" observedRunningTime="2026-02-17 00:08:36.801094501 +0000 UTC m=+190.536807262" watchObservedRunningTime="2026-02-17 00:08:36.802199544 +0000 UTC m=+190.537912295" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.816665 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2tkvq" podStartSLOduration=3.083370221 podStartE2EDuration="44.816645984s" podCreationTimestamp="2026-02-17 00:07:52 +0000 UTC" firstStartedPulling="2026-02-17 00:07:54.711630605 +0000 UTC m=+148.447343356" lastFinishedPulling="2026-02-17 00:08:36.444906368 +0000 UTC m=+190.180619119" observedRunningTime="2026-02-17 00:08:36.815535461 +0000 UTC m=+190.551248212" watchObservedRunningTime="2026-02-17 00:08:36.816645984 +0000 UTC m=+190.552358735" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.888226 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7423f424-b2b9-4045-8dce-9417d3ad6eab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7423f424-b2b9-4045-8dce-9417d3ad6eab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.888520 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7423f424-b2b9-4045-8dce-9417d3ad6eab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7423f424-b2b9-4045-8dce-9417d3ad6eab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.888547 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7423f424-b2b9-4045-8dce-9417d3ad6eab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7423f424-b2b9-4045-8dce-9417d3ad6eab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 00:08:36 crc kubenswrapper[4690]: I0217 00:08:36.915604 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7423f424-b2b9-4045-8dce-9417d3ad6eab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7423f424-b2b9-4045-8dce-9417d3ad6eab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 00:08:37 crc kubenswrapper[4690]: I0217 00:08:37.123821 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 00:08:37 crc kubenswrapper[4690]: I0217 00:08:37.584461 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 17 00:08:37 crc kubenswrapper[4690]: I0217 00:08:37.781472 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7423f424-b2b9-4045-8dce-9417d3ad6eab","Type":"ContainerStarted","Data":"2039ec249f9e81331889677c5c41d778a804520a8790432d534371f7e7091cbe"} Feb 17 00:08:37 crc kubenswrapper[4690]: I0217 00:08:37.783309 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xntcs" event={"ID":"e5df6322-f224-4e0e-8099-02865f10660a","Type":"ContainerStarted","Data":"397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b"} Feb 17 00:08:37 crc kubenswrapper[4690]: I0217 00:08:37.786341 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k62p2" event={"ID":"abbde255-fc8e-4574-b9b2-77e540002308","Type":"ContainerStarted","Data":"59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101"} Feb 17 00:08:37 crc kubenswrapper[4690]: I0217 00:08:37.788296 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbdwr" event={"ID":"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd","Type":"ContainerStarted","Data":"7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65"} Feb 17 00:08:37 crc kubenswrapper[4690]: I0217 00:08:37.807576 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xntcs" podStartSLOduration=4.327471416 podStartE2EDuration="42.807557922s" podCreationTimestamp="2026-02-17 00:07:55 +0000 UTC" firstStartedPulling="2026-02-17 00:07:58.257906609 +0000 UTC m=+151.993619360" lastFinishedPulling="2026-02-17 00:08:36.737993115 +0000 UTC m=+190.473705866" observedRunningTime="2026-02-17 00:08:37.804896913 +0000 UTC m=+191.540609664" watchObservedRunningTime="2026-02-17 00:08:37.807557922 +0000 UTC m=+191.543270673" Feb 17 00:08:37 crc kubenswrapper[4690]: I0217 00:08:37.822074 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pbdwr" podStartSLOduration=4.983776066 podStartE2EDuration="45.822056913s" podCreationTimestamp="2026-02-17 00:07:52 +0000 UTC" firstStartedPulling="2026-02-17 00:07:55.966462223 +0000 UTC m=+149.702174974" lastFinishedPulling="2026-02-17 00:08:36.80474307 +0000 UTC m=+190.540455821" observedRunningTime="2026-02-17 00:08:37.821755344 +0000 UTC m=+191.557468095" watchObservedRunningTime="2026-02-17 00:08:37.822056913 +0000 UTC m=+191.557769664" Feb 17 00:08:37 crc kubenswrapper[4690]: I0217 00:08:37.859713 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k62p2" podStartSLOduration=3.930614675 podStartE2EDuration="45.859695472s" podCreationTimestamp="2026-02-17 00:07:52 +0000 UTC" firstStartedPulling="2026-02-17 00:07:54.712411168 +0000 UTC m=+148.448123929" lastFinishedPulling="2026-02-17 00:08:36.641491975 +0000 UTC m=+190.377204726" observedRunningTime="2026-02-17 00:08:37.849775077 +0000 UTC m=+191.585487828" watchObservedRunningTime="2026-02-17 00:08:37.859695472 +0000 UTC m=+191.595408223" Feb 17 00:08:38 crc kubenswrapper[4690]: I0217 00:08:38.800323 4690 generic.go:334] "Generic (PLEG): container finished" podID="7423f424-b2b9-4045-8dce-9417d3ad6eab" containerID="75c40f80a2740b628317851ed358acb2e69ba11ccdced898141aaff5824f3efa" exitCode=0 Feb 17 00:08:38 crc kubenswrapper[4690]: I0217 00:08:38.800507 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7423f424-b2b9-4045-8dce-9417d3ad6eab","Type":"ContainerDied","Data":"75c40f80a2740b628317851ed358acb2e69ba11ccdced898141aaff5824f3efa"} Feb 17 00:08:40 crc kubenswrapper[4690]: I0217 00:08:40.118506 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 00:08:40 crc kubenswrapper[4690]: I0217 00:08:40.228696 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7423f424-b2b9-4045-8dce-9417d3ad6eab-kube-api-access\") pod \"7423f424-b2b9-4045-8dce-9417d3ad6eab\" (UID: \"7423f424-b2b9-4045-8dce-9417d3ad6eab\") " Feb 17 00:08:40 crc kubenswrapper[4690]: I0217 00:08:40.228799 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7423f424-b2b9-4045-8dce-9417d3ad6eab-kubelet-dir\") pod \"7423f424-b2b9-4045-8dce-9417d3ad6eab\" (UID: \"7423f424-b2b9-4045-8dce-9417d3ad6eab\") " Feb 17 00:08:40 crc kubenswrapper[4690]: I0217 00:08:40.228948 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7423f424-b2b9-4045-8dce-9417d3ad6eab-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7423f424-b2b9-4045-8dce-9417d3ad6eab" (UID: "7423f424-b2b9-4045-8dce-9417d3ad6eab"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:08:40 crc kubenswrapper[4690]: I0217 00:08:40.229228 4690 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7423f424-b2b9-4045-8dce-9417d3ad6eab-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:40 crc kubenswrapper[4690]: I0217 00:08:40.244079 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7423f424-b2b9-4045-8dce-9417d3ad6eab-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7423f424-b2b9-4045-8dce-9417d3ad6eab" (UID: "7423f424-b2b9-4045-8dce-9417d3ad6eab"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:08:40 crc kubenswrapper[4690]: I0217 00:08:40.333215 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7423f424-b2b9-4045-8dce-9417d3ad6eab-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:40 crc kubenswrapper[4690]: I0217 00:08:40.811569 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7423f424-b2b9-4045-8dce-9417d3ad6eab","Type":"ContainerDied","Data":"2039ec249f9e81331889677c5c41d778a804520a8790432d534371f7e7091cbe"} Feb 17 00:08:40 crc kubenswrapper[4690]: I0217 00:08:40.811609 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2039ec249f9e81331889677c5c41d778a804520a8790432d534371f7e7091cbe" Feb 17 00:08:40 crc kubenswrapper[4690]: I0217 00:08:40.811637 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 00:08:42 crc kubenswrapper[4690]: I0217 00:08:42.531857 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:08:42 crc kubenswrapper[4690]: I0217 00:08:42.532110 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:08:42 crc kubenswrapper[4690]: I0217 00:08:42.576194 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:08:42 crc kubenswrapper[4690]: I0217 00:08:42.738227 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:08:42 crc kubenswrapper[4690]: I0217 00:08:42.738273 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:08:42 crc kubenswrapper[4690]: I0217 00:08:42.781007 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:08:42 crc kubenswrapper[4690]: I0217 00:08:42.868131 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:08:42 crc kubenswrapper[4690]: I0217 00:08:42.885034 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.057598 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.057666 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.110858 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.172652 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.193093 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.256990 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.281036 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 17 00:08:43 crc kubenswrapper[4690]: E0217 00:08:43.281321 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7423f424-b2b9-4045-8dce-9417d3ad6eab" containerName="pruner" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.281344 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7423f424-b2b9-4045-8dce-9417d3ad6eab" containerName="pruner" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.281541 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7423f424-b2b9-4045-8dce-9417d3ad6eab" containerName="pruner" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.282070 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.289483 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.290373 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.293456 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.371088 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-var-lock\") pod \"installer-9-crc\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.371550 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kube-api-access\") pod \"installer-9-crc\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.371670 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.472635 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kube-api-access\") pod \"installer-9-crc\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.472704 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.472742 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-var-lock\") pod \"installer-9-crc\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.472828 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.472859 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-var-lock\") pod \"installer-9-crc\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.513292 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kube-api-access\") pod \"installer-9-crc\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.598711 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.862750 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:08:43 crc kubenswrapper[4690]: I0217 00:08:43.882719 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:08:44 crc kubenswrapper[4690]: I0217 00:08:44.053469 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 17 00:08:44 crc kubenswrapper[4690]: W0217 00:08:44.066228 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd36eab40_817c_4939_a0d6_ea1fbb2949ca.slice/crio-25beef08c99c361d59a6124afdb6d666cb1f6220d06dc7ad90f71d69857334ba WatchSource:0}: Error finding container 25beef08c99c361d59a6124afdb6d666cb1f6220d06dc7ad90f71d69857334ba: Status 404 returned error can't find the container with id 25beef08c99c361d59a6124afdb6d666cb1f6220d06dc7ad90f71d69857334ba Feb 17 00:08:44 crc kubenswrapper[4690]: I0217 00:08:44.835943 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d36eab40-817c-4939-a0d6-ea1fbb2949ca","Type":"ContainerStarted","Data":"f0d770eb77c3a1d0c44dc40b6327a87813001b1e1c5d84d96eaa1d544d3eaa9e"} Feb 17 00:08:44 crc kubenswrapper[4690]: I0217 00:08:44.836218 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d36eab40-817c-4939-a0d6-ea1fbb2949ca","Type":"ContainerStarted","Data":"25beef08c99c361d59a6124afdb6d666cb1f6220d06dc7ad90f71d69857334ba"} Feb 17 00:08:44 crc kubenswrapper[4690]: I0217 00:08:44.857060 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.857039944 podStartE2EDuration="1.857039944s" podCreationTimestamp="2026-02-17 00:08:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:08:44.855908951 +0000 UTC m=+198.591621732" watchObservedRunningTime="2026-02-17 00:08:44.857039944 +0000 UTC m=+198.592752695" Feb 17 00:08:44 crc kubenswrapper[4690]: I0217 00:08:44.883692 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k62p2"] Feb 17 00:08:45 crc kubenswrapper[4690]: I0217 00:08:45.164675 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:08:45 crc kubenswrapper[4690]: I0217 00:08:45.841399 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k62p2" podUID="abbde255-fc8e-4574-b9b2-77e540002308" containerName="registry-server" containerID="cri-o://59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101" gracePeriod=2 Feb 17 00:08:45 crc kubenswrapper[4690]: I0217 00:08:45.999318 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:08:45 crc kubenswrapper[4690]: I0217 00:08:45.999385 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.070674 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.274155 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.367013 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.367048 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.413389 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.414334 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vdg7\" (UniqueName: \"kubernetes.io/projected/abbde255-fc8e-4574-b9b2-77e540002308-kube-api-access-8vdg7\") pod \"abbde255-fc8e-4574-b9b2-77e540002308\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.414684 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-utilities\") pod \"abbde255-fc8e-4574-b9b2-77e540002308\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.415224 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-catalog-content\") pod \"abbde255-fc8e-4574-b9b2-77e540002308\" (UID: \"abbde255-fc8e-4574-b9b2-77e540002308\") " Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.416083 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-utilities" (OuterVolumeSpecName: "utilities") pod "abbde255-fc8e-4574-b9b2-77e540002308" (UID: "abbde255-fc8e-4574-b9b2-77e540002308"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.432968 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abbde255-fc8e-4574-b9b2-77e540002308-kube-api-access-8vdg7" (OuterVolumeSpecName: "kube-api-access-8vdg7") pod "abbde255-fc8e-4574-b9b2-77e540002308" (UID: "abbde255-fc8e-4574-b9b2-77e540002308"). InnerVolumeSpecName "kube-api-access-8vdg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.508199 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abbde255-fc8e-4574-b9b2-77e540002308" (UID: "abbde255-fc8e-4574-b9b2-77e540002308"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.517048 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vdg7\" (UniqueName: \"kubernetes.io/projected/abbde255-fc8e-4574-b9b2-77e540002308-kube-api-access-8vdg7\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.517106 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.517116 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abbde255-fc8e-4574-b9b2-77e540002308-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.852671 4690 generic.go:334] "Generic (PLEG): container finished" podID="abbde255-fc8e-4574-b9b2-77e540002308" containerID="59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101" exitCode=0 Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.852799 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k62p2" event={"ID":"abbde255-fc8e-4574-b9b2-77e540002308","Type":"ContainerDied","Data":"59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101"} Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.852865 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k62p2" event={"ID":"abbde255-fc8e-4574-b9b2-77e540002308","Type":"ContainerDied","Data":"f4c2b952dba4dacd3437574ca8b7734676f39447ce23a06f6ace7877d4a0fe37"} Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.852905 4690 scope.go:117] "RemoveContainer" containerID="59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.854103 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k62p2" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.874521 4690 scope.go:117] "RemoveContainer" containerID="aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.901018 4690 scope.go:117] "RemoveContainer" containerID="de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.908390 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.931940 4690 scope.go:117] "RemoveContainer" containerID="59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101" Feb 17 00:08:46 crc kubenswrapper[4690]: E0217 00:08:46.932810 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101\": container with ID starting with 59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101 not found: ID does not exist" containerID="59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.932988 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101"} err="failed to get container status \"59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101\": rpc error: code = NotFound desc = could not find container \"59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101\": container with ID starting with 59279c867193da7ad6cc5e76555821bcee254002befe691fbad8f8595fb90101 not found: ID does not exist" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.933189 4690 scope.go:117] "RemoveContainer" containerID="aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9" Feb 17 00:08:46 crc kubenswrapper[4690]: E0217 00:08:46.935003 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9\": container with ID starting with aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9 not found: ID does not exist" containerID="aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.935073 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9"} err="failed to get container status \"aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9\": rpc error: code = NotFound desc = could not find container \"aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9\": container with ID starting with aa80c3959b37bc402f64313cc2553f5225631661f61db2dd1d68ea823edbb6d9 not found: ID does not exist" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.935115 4690 scope.go:117] "RemoveContainer" containerID="de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37" Feb 17 00:08:46 crc kubenswrapper[4690]: E0217 00:08:46.935519 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37\": container with ID starting with de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37 not found: ID does not exist" containerID="de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.935548 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37"} err="failed to get container status \"de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37\": rpc error: code = NotFound desc = could not find container \"de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37\": container with ID starting with de14762bd60773a06102476b60d0015fd9a19ab9f175e493d8979afe359a9f37 not found: ID does not exist" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.936474 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.947679 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k62p2"] Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.952615 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k62p2"] Feb 17 00:08:46 crc kubenswrapper[4690]: I0217 00:08:46.986199 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abbde255-fc8e-4574-b9b2-77e540002308" path="/var/lib/kubelet/pods/abbde255-fc8e-4574-b9b2-77e540002308/volumes" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.087585 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-grqqc"] Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.087882 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-grqqc" podUID="573d80ca-54d0-4c1a-8987-e734917e9900" containerName="registry-server" containerID="cri-o://091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34" gracePeriod=2 Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.292187 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pbdwr"] Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.292541 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pbdwr" podUID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerName="registry-server" containerID="cri-o://7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65" gracePeriod=2 Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.558316 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.635929 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t57sb\" (UniqueName: \"kubernetes.io/projected/573d80ca-54d0-4c1a-8987-e734917e9900-kube-api-access-t57sb\") pod \"573d80ca-54d0-4c1a-8987-e734917e9900\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.636089 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-utilities\") pod \"573d80ca-54d0-4c1a-8987-e734917e9900\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.636145 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-catalog-content\") pod \"573d80ca-54d0-4c1a-8987-e734917e9900\" (UID: \"573d80ca-54d0-4c1a-8987-e734917e9900\") " Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.636793 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-utilities" (OuterVolumeSpecName: "utilities") pod "573d80ca-54d0-4c1a-8987-e734917e9900" (UID: "573d80ca-54d0-4c1a-8987-e734917e9900"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.639099 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/573d80ca-54d0-4c1a-8987-e734917e9900-kube-api-access-t57sb" (OuterVolumeSpecName: "kube-api-access-t57sb") pod "573d80ca-54d0-4c1a-8987-e734917e9900" (UID: "573d80ca-54d0-4c1a-8987-e734917e9900"). InnerVolumeSpecName "kube-api-access-t57sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.650086 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.669693 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "573d80ca-54d0-4c1a-8987-e734917e9900" (UID: "573d80ca-54d0-4c1a-8987-e734917e9900"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.737330 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq88l\" (UniqueName: \"kubernetes.io/projected/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-kube-api-access-cq88l\") pod \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.737482 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-catalog-content\") pod \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.737523 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-utilities\") pod \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\" (UID: \"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd\") " Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.738508 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.738549 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/573d80ca-54d0-4c1a-8987-e734917e9900-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.738571 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t57sb\" (UniqueName: \"kubernetes.io/projected/573d80ca-54d0-4c1a-8987-e734917e9900-kube-api-access-t57sb\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.738671 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-utilities" (OuterVolumeSpecName: "utilities") pod "a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" (UID: "a24db2e5-dbd4-40be-bf3b-4ec15f5329bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.741802 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-kube-api-access-cq88l" (OuterVolumeSpecName: "kube-api-access-cq88l") pod "a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" (UID: "a24db2e5-dbd4-40be-bf3b-4ec15f5329bd"). InnerVolumeSpecName "kube-api-access-cq88l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.808831 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" (UID: "a24db2e5-dbd4-40be-bf3b-4ec15f5329bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.839634 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq88l\" (UniqueName: \"kubernetes.io/projected/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-kube-api-access-cq88l\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.839682 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.839695 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.861768 4690 generic.go:334] "Generic (PLEG): container finished" podID="573d80ca-54d0-4c1a-8987-e734917e9900" containerID="091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34" exitCode=0 Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.861831 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grqqc" event={"ID":"573d80ca-54d0-4c1a-8987-e734917e9900","Type":"ContainerDied","Data":"091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34"} Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.861858 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grqqc" event={"ID":"573d80ca-54d0-4c1a-8987-e734917e9900","Type":"ContainerDied","Data":"67607fd6c1d2383b709f039d3cb16d20c24337b8029257f7ebc496017f01cee0"} Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.861861 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grqqc" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.861875 4690 scope.go:117] "RemoveContainer" containerID="091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.866280 4690 generic.go:334] "Generic (PLEG): container finished" podID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerID="7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65" exitCode=0 Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.866355 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbdwr" event={"ID":"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd","Type":"ContainerDied","Data":"7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65"} Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.866458 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbdwr" event={"ID":"a24db2e5-dbd4-40be-bf3b-4ec15f5329bd","Type":"ContainerDied","Data":"56ff16172e3dcc5abc967165749332600dd25fb8657186e8fc2d7b77695f6e82"} Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.866581 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pbdwr" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.881508 4690 scope.go:117] "RemoveContainer" containerID="b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.909101 4690 scope.go:117] "RemoveContainer" containerID="da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.913450 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-grqqc"] Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.922581 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-grqqc"] Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.926277 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pbdwr"] Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.928864 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pbdwr"] Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.934593 4690 scope.go:117] "RemoveContainer" containerID="091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34" Feb 17 00:08:47 crc kubenswrapper[4690]: E0217 00:08:47.935169 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34\": container with ID starting with 091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34 not found: ID does not exist" containerID="091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.935217 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34"} err="failed to get container status \"091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34\": rpc error: code = NotFound desc = could not find container \"091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34\": container with ID starting with 091c3b82efa7ccca6cb5cc1a034fab26ce4877a80cf3dfb0d3789e2e5a773a34 not found: ID does not exist" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.935253 4690 scope.go:117] "RemoveContainer" containerID="b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459" Feb 17 00:08:47 crc kubenswrapper[4690]: E0217 00:08:47.935678 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459\": container with ID starting with b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459 not found: ID does not exist" containerID="b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.935806 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459"} err="failed to get container status \"b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459\": rpc error: code = NotFound desc = could not find container \"b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459\": container with ID starting with b54f9d26971dd39dae2d00f59ffc2a0e5378f0812118caa004e2199cd2572459 not found: ID does not exist" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.935917 4690 scope.go:117] "RemoveContainer" containerID="da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092" Feb 17 00:08:47 crc kubenswrapper[4690]: E0217 00:08:47.936555 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092\": container with ID starting with da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092 not found: ID does not exist" containerID="da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.936598 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092"} err="failed to get container status \"da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092\": rpc error: code = NotFound desc = could not find container \"da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092\": container with ID starting with da1f530dddcfa0551b3b6036bce4a2593354c88fbe97247ee1f44ee900e41092 not found: ID does not exist" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.936623 4690 scope.go:117] "RemoveContainer" containerID="7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65" Feb 17 00:08:47 crc kubenswrapper[4690]: I0217 00:08:47.953037 4690 scope.go:117] "RemoveContainer" containerID="c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec" Feb 17 00:08:48 crc kubenswrapper[4690]: I0217 00:08:48.010484 4690 scope.go:117] "RemoveContainer" containerID="ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae" Feb 17 00:08:48 crc kubenswrapper[4690]: I0217 00:08:48.031584 4690 scope.go:117] "RemoveContainer" containerID="7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65" Feb 17 00:08:48 crc kubenswrapper[4690]: E0217 00:08:48.032060 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65\": container with ID starting with 7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65 not found: ID does not exist" containerID="7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65" Feb 17 00:08:48 crc kubenswrapper[4690]: I0217 00:08:48.032191 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65"} err="failed to get container status \"7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65\": rpc error: code = NotFound desc = could not find container \"7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65\": container with ID starting with 7d143f9b0e434bdf7ac46d4bf91e5678637685e9b42f98d45a48f5e13edebe65 not found: ID does not exist" Feb 17 00:08:48 crc kubenswrapper[4690]: I0217 00:08:48.032313 4690 scope.go:117] "RemoveContainer" containerID="c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec" Feb 17 00:08:48 crc kubenswrapper[4690]: E0217 00:08:48.032761 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec\": container with ID starting with c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec not found: ID does not exist" containerID="c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec" Feb 17 00:08:48 crc kubenswrapper[4690]: I0217 00:08:48.032815 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec"} err="failed to get container status \"c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec\": rpc error: code = NotFound desc = could not find container \"c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec\": container with ID starting with c393e6dd2fa6d5e4053269c4e7c541740138bceaffcf16eeffda97b7758d0cec not found: ID does not exist" Feb 17 00:08:48 crc kubenswrapper[4690]: I0217 00:08:48.032852 4690 scope.go:117] "RemoveContainer" containerID="ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae" Feb 17 00:08:48 crc kubenswrapper[4690]: E0217 00:08:48.033288 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae\": container with ID starting with ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae not found: ID does not exist" containerID="ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae" Feb 17 00:08:48 crc kubenswrapper[4690]: I0217 00:08:48.033408 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae"} err="failed to get container status \"ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae\": rpc error: code = NotFound desc = could not find container \"ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae\": container with ID starting with ac521afa9fbafd956d17eec234f78996bb0e4e29a54fa18eadd2096e3b0cbaae not found: ID does not exist" Feb 17 00:08:48 crc kubenswrapper[4690]: I0217 00:08:48.987793 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="573d80ca-54d0-4c1a-8987-e734917e9900" path="/var/lib/kubelet/pods/573d80ca-54d0-4c1a-8987-e734917e9900/volumes" Feb 17 00:08:48 crc kubenswrapper[4690]: I0217 00:08:48.989138 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" path="/var/lib/kubelet/pods/a24db2e5-dbd4-40be-bf3b-4ec15f5329bd/volumes" Feb 17 00:08:49 crc kubenswrapper[4690]: I0217 00:08:49.682122 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zd95f"] Feb 17 00:08:49 crc kubenswrapper[4690]: I0217 00:08:49.682551 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zd95f" podUID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerName="registry-server" containerID="cri-o://f8feb44d671255344b6c5e4472af6be49efaa0f20aadd8771a9cec6866973bf7" gracePeriod=2 Feb 17 00:08:49 crc kubenswrapper[4690]: I0217 00:08:49.884267 4690 generic.go:334] "Generic (PLEG): container finished" podID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerID="f8feb44d671255344b6c5e4472af6be49efaa0f20aadd8771a9cec6866973bf7" exitCode=0 Feb 17 00:08:49 crc kubenswrapper[4690]: I0217 00:08:49.884325 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd95f" event={"ID":"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2","Type":"ContainerDied","Data":"f8feb44d671255344b6c5e4472af6be49efaa0f20aadd8771a9cec6866973bf7"} Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.065246 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.172189 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x528p\" (UniqueName: \"kubernetes.io/projected/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-kube-api-access-x528p\") pod \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.172696 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-utilities\") pod \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.172754 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-catalog-content\") pod \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\" (UID: \"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2\") " Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.173768 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-utilities" (OuterVolumeSpecName: "utilities") pod "7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" (UID: "7fc6425a-4fc1-4084-bbc1-8e69f6545fd2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.181606 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-kube-api-access-x528p" (OuterVolumeSpecName: "kube-api-access-x528p") pod "7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" (UID: "7fc6425a-4fc1-4084-bbc1-8e69f6545fd2"). InnerVolumeSpecName "kube-api-access-x528p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.274202 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x528p\" (UniqueName: \"kubernetes.io/projected/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-kube-api-access-x528p\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.274599 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.346385 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" (UID: "7fc6425a-4fc1-4084-bbc1-8e69f6545fd2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.375684 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.860953 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" podUID="b7b8197f-4d82-4a0b-bed6-485dfaab9024" containerName="oauth-openshift" containerID="cri-o://4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490" gracePeriod=15 Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.894500 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd95f" event={"ID":"7fc6425a-4fc1-4084-bbc1-8e69f6545fd2","Type":"ContainerDied","Data":"05e215f7b413970df290bee3e1ade448f1dd264cc6b8a68bc7ebb78582889507"} Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.894820 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zd95f" Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.894874 4690 scope.go:117] "RemoveContainer" containerID="f8feb44d671255344b6c5e4472af6be49efaa0f20aadd8771a9cec6866973bf7" Feb 17 00:08:50 crc kubenswrapper[4690]: I0217 00:08:50.923906 4690 scope.go:117] "RemoveContainer" containerID="8266250e18fe1fc635cde0f323106e87f3250831d64b4c0886160dfc1f60bae7" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.000321 4690 scope.go:117] "RemoveContainer" containerID="f0b15c526fdeec1ff5cd1c5f04abd6c4010b4df1e9788184b131bdb37d4d78a5" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.023018 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zd95f"] Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.026414 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zd95f"] Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.235213 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387463 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-policies\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387551 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-session\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387590 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-provider-selection\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387618 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-service-ca\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387670 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-dir\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387702 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-ocp-branding-template\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387725 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-login\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387754 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f6jp\" (UniqueName: \"kubernetes.io/projected/b7b8197f-4d82-4a0b-bed6-485dfaab9024-kube-api-access-2f6jp\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387778 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-cliconfig\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387820 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-serving-cert\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387852 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-idp-0-file-data\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387893 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-trusted-ca-bundle\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387925 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-error\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.387952 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-router-certs\") pod \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\" (UID: \"b7b8197f-4d82-4a0b-bed6-485dfaab9024\") " Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.388049 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.388328 4690 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.388405 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.388550 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.388807 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.388973 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.392537 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.392959 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.393078 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.393574 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7b8197f-4d82-4a0b-bed6-485dfaab9024-kube-api-access-2f6jp" (OuterVolumeSpecName: "kube-api-access-2f6jp") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "kube-api-access-2f6jp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.393779 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.393847 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.393961 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.394091 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.396099 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "b7b8197f-4d82-4a0b-bed6-485dfaab9024" (UID: "b7b8197f-4d82-4a0b-bed6-485dfaab9024"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489505 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489568 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489585 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f6jp\" (UniqueName: \"kubernetes.io/projected/b7b8197f-4d82-4a0b-bed6-485dfaab9024-kube-api-access-2f6jp\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489602 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489618 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489631 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489644 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489659 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489674 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489687 4690 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489715 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489730 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.489742 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7b8197f-4d82-4a0b-bed6-485dfaab9024-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.902194 4690 generic.go:334] "Generic (PLEG): container finished" podID="b7b8197f-4d82-4a0b-bed6-485dfaab9024" containerID="4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490" exitCode=0 Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.902298 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" event={"ID":"b7b8197f-4d82-4a0b-bed6-485dfaab9024","Type":"ContainerDied","Data":"4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490"} Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.902352 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" event={"ID":"b7b8197f-4d82-4a0b-bed6-485dfaab9024","Type":"ContainerDied","Data":"dc33dc5fc7af820e3e7edc572c362cba708bcf944b49af38c0a57cd91cd475c2"} Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.902391 4690 scope.go:117] "RemoveContainer" containerID="4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.902404 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-c6pc7" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.924932 4690 scope.go:117] "RemoveContainer" containerID="4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490" Feb 17 00:08:51 crc kubenswrapper[4690]: E0217 00:08:51.927088 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490\": container with ID starting with 4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490 not found: ID does not exist" containerID="4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.927134 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490"} err="failed to get container status \"4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490\": rpc error: code = NotFound desc = could not find container \"4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490\": container with ID starting with 4bc89539aa027c253178dfbb5a30bcc22c0a800be2d676b1d2635a44f863a490 not found: ID does not exist" Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.947417 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-c6pc7"] Feb 17 00:08:51 crc kubenswrapper[4690]: I0217 00:08:51.955349 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-c6pc7"] Feb 17 00:08:52 crc kubenswrapper[4690]: I0217 00:08:52.989012 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" path="/var/lib/kubelet/pods/7fc6425a-4fc1-4084-bbc1-8e69f6545fd2/volumes" Feb 17 00:08:52 crc kubenswrapper[4690]: I0217 00:08:52.990700 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7b8197f-4d82-4a0b-bed6-485dfaab9024" path="/var/lib/kubelet/pods/b7b8197f-4d82-4a0b-bed6-485dfaab9024/volumes" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.176329 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-575cc5b957-tm5qq"] Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177057 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerName="extract-utilities" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177082 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerName="extract-utilities" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177098 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="573d80ca-54d0-4c1a-8987-e734917e9900" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177112 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="573d80ca-54d0-4c1a-8987-e734917e9900" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177130 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="573d80ca-54d0-4c1a-8987-e734917e9900" containerName="extract-content" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177142 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="573d80ca-54d0-4c1a-8987-e734917e9900" containerName="extract-content" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177163 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177175 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177197 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abbde255-fc8e-4574-b9b2-77e540002308" containerName="extract-utilities" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177209 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="abbde255-fc8e-4574-b9b2-77e540002308" containerName="extract-utilities" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177228 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="573d80ca-54d0-4c1a-8987-e734917e9900" containerName="extract-utilities" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177240 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="573d80ca-54d0-4c1a-8987-e734917e9900" containerName="extract-utilities" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177263 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b8197f-4d82-4a0b-bed6-485dfaab9024" containerName="oauth-openshift" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177276 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b8197f-4d82-4a0b-bed6-485dfaab9024" containerName="oauth-openshift" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177292 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abbde255-fc8e-4574-b9b2-77e540002308" containerName="extract-content" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177303 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="abbde255-fc8e-4574-b9b2-77e540002308" containerName="extract-content" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177503 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerName="extract-content" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177522 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerName="extract-content" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177541 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abbde255-fc8e-4574-b9b2-77e540002308" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177555 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="abbde255-fc8e-4574-b9b2-77e540002308" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177594 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerName="extract-content" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177618 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerName="extract-content" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177637 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerName="extract-utilities" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177653 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerName="extract-utilities" Feb 17 00:08:55 crc kubenswrapper[4690]: E0217 00:08:55.177676 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177691 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177907 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a24db2e5-dbd4-40be-bf3b-4ec15f5329bd" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177933 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7b8197f-4d82-4a0b-bed6-485dfaab9024" containerName="oauth-openshift" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177958 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="573d80ca-54d0-4c1a-8987-e734917e9900" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.177991 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="abbde255-fc8e-4574-b9b2-77e540002308" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.178017 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc6425a-4fc1-4084-bbc1-8e69f6545fd2" containerName="registry-server" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.178656 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.181876 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.183387 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.183638 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.183674 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.186856 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.187119 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.187315 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.187654 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.187781 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.188760 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.189173 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.189965 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.199179 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-575cc5b957-tm5qq"] Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.208728 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.213979 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.217332 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.241622 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-audit-policies\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.241702 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8mkm\" (UniqueName: \"kubernetes.io/projected/ae24c24f-523f-47ef-abff-46932ff577db-kube-api-access-c8mkm\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.241761 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-template-login\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.241834 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-service-ca\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.241882 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-session\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.241918 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.241971 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.242010 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-cliconfig\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.242086 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae24c24f-523f-47ef-abff-46932ff577db-audit-dir\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.242122 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-router-certs\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.242163 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.242201 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.242237 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-template-error\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.242310 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-serving-cert\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.344230 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-template-login\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.344760 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-service-ca\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.345027 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-session\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.345211 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.345441 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.345606 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-cliconfig\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.345858 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae24c24f-523f-47ef-abff-46932ff577db-audit-dir\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.346235 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-router-certs\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.346156 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae24c24f-523f-47ef-abff-46932ff577db-audit-dir\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.346310 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-service-ca\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.347128 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-cliconfig\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.347287 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.347737 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.348058 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.348255 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-template-error\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.348496 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-serving-cert\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.348664 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-audit-policies\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.348834 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8mkm\" (UniqueName: \"kubernetes.io/projected/ae24c24f-523f-47ef-abff-46932ff577db-kube-api-access-c8mkm\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.350320 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae24c24f-523f-47ef-abff-46932ff577db-audit-policies\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.351863 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.352764 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-router-certs\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.352772 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-template-login\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.353709 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-session\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.361734 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-system-serving-cert\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.361802 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-template-error\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.362719 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.363126 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae24c24f-523f-47ef-abff-46932ff577db-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.384317 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8mkm\" (UniqueName: \"kubernetes.io/projected/ae24c24f-523f-47ef-abff-46932ff577db-kube-api-access-c8mkm\") pod \"oauth-openshift-575cc5b957-tm5qq\" (UID: \"ae24c24f-523f-47ef-abff-46932ff577db\") " pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:55 crc kubenswrapper[4690]: I0217 00:08:55.514040 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.032902 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-575cc5b957-tm5qq"] Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.325867 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.326481 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.326554 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.327455 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6"} pod="openshift-machine-config-operator/machine-config-daemon-lszwm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.327599 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" containerID="cri-o://e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6" gracePeriod=600 Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.942315 4690 generic.go:334] "Generic (PLEG): container finished" podID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerID="e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6" exitCode=0 Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.942571 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerDied","Data":"e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6"} Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.942964 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerStarted","Data":"994b411b25ff62fc16205928af96a4f9d2be45f24e3023cf34d1411551126d69"} Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.947200 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" event={"ID":"ae24c24f-523f-47ef-abff-46932ff577db","Type":"ContainerStarted","Data":"76f8627b33b6844c82e9829cd87c453cec0b08b4540aa8f263e7050001e1527e"} Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.947252 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" event={"ID":"ae24c24f-523f-47ef-abff-46932ff577db","Type":"ContainerStarted","Data":"344403f58cbffca9b4b3d7f5c567e8e28090d154567614958937facb693ef4b4"} Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.948351 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:56 crc kubenswrapper[4690]: I0217 00:08:56.958868 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" Feb 17 00:08:57 crc kubenswrapper[4690]: I0217 00:08:57.009549 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-575cc5b957-tm5qq" podStartSLOduration=32.009515381 podStartE2EDuration="32.009515381s" podCreationTimestamp="2026-02-17 00:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:08:57.007220363 +0000 UTC m=+210.742933154" watchObservedRunningTime="2026-02-17 00:08:57.009515381 +0000 UTC m=+210.745228172" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.063470 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2tkvq"] Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.064268 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2tkvq" podUID="c08acd43-80f4-4a48-b848-7899561e26b5" containerName="registry-server" containerID="cri-o://7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2" gracePeriod=30 Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.074601 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fcmdk"] Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.075129 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fcmdk" podUID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerName="registry-server" containerID="cri-o://5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec" gracePeriod=30 Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.086592 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gxcpk"] Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.086792 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" podUID="4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" containerName="marketplace-operator" containerID="cri-o://9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f" gracePeriod=30 Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.103205 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pv76l"] Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.103540 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pv76l" podUID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerName="registry-server" containerID="cri-o://8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a" gracePeriod=30 Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.111265 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xntcs"] Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.111674 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xntcs" podUID="e5df6322-f224-4e0e-8099-02865f10660a" containerName="registry-server" containerID="cri-o://397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b" gracePeriod=30 Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.119268 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvvgs"] Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.120247 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.124391 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvvgs"] Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.200311 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75scc\" (UniqueName: \"kubernetes.io/projected/4372c018-0aaf-459e-9e8f-6423d7ed5695-kube-api-access-75scc\") pod \"marketplace-operator-79b997595-wvvgs\" (UID: \"4372c018-0aaf-459e-9e8f-6423d7ed5695\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.200401 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4372c018-0aaf-459e-9e8f-6423d7ed5695-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wvvgs\" (UID: \"4372c018-0aaf-459e-9e8f-6423d7ed5695\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.200446 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4372c018-0aaf-459e-9e8f-6423d7ed5695-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wvvgs\" (UID: \"4372c018-0aaf-459e-9e8f-6423d7ed5695\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.301241 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75scc\" (UniqueName: \"kubernetes.io/projected/4372c018-0aaf-459e-9e8f-6423d7ed5695-kube-api-access-75scc\") pod \"marketplace-operator-79b997595-wvvgs\" (UID: \"4372c018-0aaf-459e-9e8f-6423d7ed5695\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.301328 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4372c018-0aaf-459e-9e8f-6423d7ed5695-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wvvgs\" (UID: \"4372c018-0aaf-459e-9e8f-6423d7ed5695\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.301393 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4372c018-0aaf-459e-9e8f-6423d7ed5695-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wvvgs\" (UID: \"4372c018-0aaf-459e-9e8f-6423d7ed5695\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.302838 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4372c018-0aaf-459e-9e8f-6423d7ed5695-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wvvgs\" (UID: \"4372c018-0aaf-459e-9e8f-6423d7ed5695\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.308224 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4372c018-0aaf-459e-9e8f-6423d7ed5695-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wvvgs\" (UID: \"4372c018-0aaf-459e-9e8f-6423d7ed5695\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.316820 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75scc\" (UniqueName: \"kubernetes.io/projected/4372c018-0aaf-459e-9e8f-6423d7ed5695-kube-api-access-75scc\") pod \"marketplace-operator-79b997595-wvvgs\" (UID: \"4372c018-0aaf-459e-9e8f-6423d7ed5695\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.433949 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.568999 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.575033 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.660090 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.668583 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.694095 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709337 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-catalog-content\") pod \"c08acd43-80f4-4a48-b848-7899561e26b5\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709386 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbkvf\" (UniqueName: \"kubernetes.io/projected/55ca888a-bbfa-477d-9769-8eab8a6aee98-kube-api-access-tbkvf\") pod \"55ca888a-bbfa-477d-9769-8eab8a6aee98\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709409 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-catalog-content\") pod \"e5df6322-f224-4e0e-8099-02865f10660a\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709428 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2nz2\" (UniqueName: \"kubernetes.io/projected/c08acd43-80f4-4a48-b848-7899561e26b5-kube-api-access-b2nz2\") pod \"c08acd43-80f4-4a48-b848-7899561e26b5\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709454 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-catalog-content\") pod \"55ca888a-bbfa-477d-9769-8eab8a6aee98\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709469 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-utilities\") pod \"e5df6322-f224-4e0e-8099-02865f10660a\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709498 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kcw7\" (UniqueName: \"kubernetes.io/projected/4c5d01e4-052c-4541-a385-79a8066bba8f-kube-api-access-7kcw7\") pod \"4c5d01e4-052c-4541-a385-79a8066bba8f\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709515 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggvrj\" (UniqueName: \"kubernetes.io/projected/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-kube-api-access-ggvrj\") pod \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709533 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-utilities\") pod \"55ca888a-bbfa-477d-9769-8eab8a6aee98\" (UID: \"55ca888a-bbfa-477d-9769-8eab8a6aee98\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709552 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-utilities\") pod \"4c5d01e4-052c-4541-a385-79a8066bba8f\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709574 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-trusted-ca\") pod \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709589 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-utilities\") pod \"c08acd43-80f4-4a48-b848-7899561e26b5\" (UID: \"c08acd43-80f4-4a48-b848-7899561e26b5\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709609 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-catalog-content\") pod \"4c5d01e4-052c-4541-a385-79a8066bba8f\" (UID: \"4c5d01e4-052c-4541-a385-79a8066bba8f\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709629 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-operator-metrics\") pod \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\" (UID: \"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.709651 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5v2z\" (UniqueName: \"kubernetes.io/projected/e5df6322-f224-4e0e-8099-02865f10660a-kube-api-access-m5v2z\") pod \"e5df6322-f224-4e0e-8099-02865f10660a\" (UID: \"e5df6322-f224-4e0e-8099-02865f10660a\") " Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.719055 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-utilities" (OuterVolumeSpecName: "utilities") pod "e5df6322-f224-4e0e-8099-02865f10660a" (UID: "e5df6322-f224-4e0e-8099-02865f10660a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.721909 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-utilities" (OuterVolumeSpecName: "utilities") pod "c08acd43-80f4-4a48-b848-7899561e26b5" (UID: "c08acd43-80f4-4a48-b848-7899561e26b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.722730 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" (UID: "4b79afc2-a76d-4fb3-8be6-ec2d3902b33c"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.722851 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-utilities" (OuterVolumeSpecName: "utilities") pod "4c5d01e4-052c-4541-a385-79a8066bba8f" (UID: "4c5d01e4-052c-4541-a385-79a8066bba8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.725609 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-kube-api-access-ggvrj" (OuterVolumeSpecName: "kube-api-access-ggvrj") pod "4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" (UID: "4b79afc2-a76d-4fb3-8be6-ec2d3902b33c"). InnerVolumeSpecName "kube-api-access-ggvrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.725765 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-utilities" (OuterVolumeSpecName: "utilities") pod "55ca888a-bbfa-477d-9769-8eab8a6aee98" (UID: "55ca888a-bbfa-477d-9769-8eab8a6aee98"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.726282 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c5d01e4-052c-4541-a385-79a8066bba8f-kube-api-access-7kcw7" (OuterVolumeSpecName: "kube-api-access-7kcw7") pod "4c5d01e4-052c-4541-a385-79a8066bba8f" (UID: "4c5d01e4-052c-4541-a385-79a8066bba8f"). InnerVolumeSpecName "kube-api-access-7kcw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.726477 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55ca888a-bbfa-477d-9769-8eab8a6aee98-kube-api-access-tbkvf" (OuterVolumeSpecName: "kube-api-access-tbkvf") pod "55ca888a-bbfa-477d-9769-8eab8a6aee98" (UID: "55ca888a-bbfa-477d-9769-8eab8a6aee98"). InnerVolumeSpecName "kube-api-access-tbkvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.727623 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" (UID: "4b79afc2-a76d-4fb3-8be6-ec2d3902b33c"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.740635 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5df6322-f224-4e0e-8099-02865f10660a-kube-api-access-m5v2z" (OuterVolumeSpecName: "kube-api-access-m5v2z") pod "e5df6322-f224-4e0e-8099-02865f10660a" (UID: "e5df6322-f224-4e0e-8099-02865f10660a"). InnerVolumeSpecName "kube-api-access-m5v2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.753527 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c08acd43-80f4-4a48-b848-7899561e26b5-kube-api-access-b2nz2" (OuterVolumeSpecName: "kube-api-access-b2nz2") pod "c08acd43-80f4-4a48-b848-7899561e26b5" (UID: "c08acd43-80f4-4a48-b848-7899561e26b5"). InnerVolumeSpecName "kube-api-access-b2nz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.773264 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvvgs"] Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.788214 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c08acd43-80f4-4a48-b848-7899561e26b5" (UID: "c08acd43-80f4-4a48-b848-7899561e26b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.794419 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55ca888a-bbfa-477d-9769-8eab8a6aee98" (UID: "55ca888a-bbfa-477d-9769-8eab8a6aee98"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812343 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812431 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kcw7\" (UniqueName: \"kubernetes.io/projected/4c5d01e4-052c-4541-a385-79a8066bba8f-kube-api-access-7kcw7\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812450 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggvrj\" (UniqueName: \"kubernetes.io/projected/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-kube-api-access-ggvrj\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812461 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812474 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812485 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812497 4690 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812511 4690 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812524 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5v2z\" (UniqueName: \"kubernetes.io/projected/e5df6322-f224-4e0e-8099-02865f10660a-kube-api-access-m5v2z\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812535 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c08acd43-80f4-4a48-b848-7899561e26b5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812546 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbkvf\" (UniqueName: \"kubernetes.io/projected/55ca888a-bbfa-477d-9769-8eab8a6aee98-kube-api-access-tbkvf\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812559 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2nz2\" (UniqueName: \"kubernetes.io/projected/c08acd43-80f4-4a48-b848-7899561e26b5-kube-api-access-b2nz2\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.812570 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55ca888a-bbfa-477d-9769-8eab8a6aee98-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.827409 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c5d01e4-052c-4541-a385-79a8066bba8f" (UID: "4c5d01e4-052c-4541-a385-79a8066bba8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.913807 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5d01e4-052c-4541-a385-79a8066bba8f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:15 crc kubenswrapper[4690]: I0217 00:09:15.944327 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5df6322-f224-4e0e-8099-02865f10660a" (UID: "e5df6322-f224-4e0e-8099-02865f10660a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.015236 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5df6322-f224-4e0e-8099-02865f10660a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.042700 4690 generic.go:334] "Generic (PLEG): container finished" podID="e5df6322-f224-4e0e-8099-02865f10660a" containerID="397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b" exitCode=0 Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.042751 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xntcs" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.042812 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xntcs" event={"ID":"e5df6322-f224-4e0e-8099-02865f10660a","Type":"ContainerDied","Data":"397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.042889 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xntcs" event={"ID":"e5df6322-f224-4e0e-8099-02865f10660a","Type":"ContainerDied","Data":"7b7ad9439d0a6c91a7feb314bf32ca6c563429bc3b3d5109067561e6ac6e56c1"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.042919 4690 scope.go:117] "RemoveContainer" containerID="397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.044469 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" event={"ID":"4372c018-0aaf-459e-9e8f-6423d7ed5695","Type":"ContainerStarted","Data":"9550af4dc5cc759223dacf2031c261d14ce2150e98ce870b965c8600dba1c785"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.044515 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" event={"ID":"4372c018-0aaf-459e-9e8f-6423d7ed5695","Type":"ContainerStarted","Data":"0edda6546bfa9800541199b227915313565fd01263408d46697fbdcfa1f9afde"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.045958 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.046763 4690 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wvvgs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.58:8080/healthz\": dial tcp 10.217.0.58:8080: connect: connection refused" start-of-body= Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.046815 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" podUID="4372c018-0aaf-459e-9e8f-6423d7ed5695" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.58:8080/healthz\": dial tcp 10.217.0.58:8080: connect: connection refused" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.047547 4690 generic.go:334] "Generic (PLEG): container finished" podID="c08acd43-80f4-4a48-b848-7899561e26b5" containerID="7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2" exitCode=0 Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.047603 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2tkvq" event={"ID":"c08acd43-80f4-4a48-b848-7899561e26b5","Type":"ContainerDied","Data":"7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.047624 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2tkvq" event={"ID":"c08acd43-80f4-4a48-b848-7899561e26b5","Type":"ContainerDied","Data":"e15ae248a29acec1068184563d7625566e9ccfb2ee038754f31a628577d5a9ca"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.047648 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2tkvq" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.057018 4690 scope.go:117] "RemoveContainer" containerID="e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.057179 4690 generic.go:334] "Generic (PLEG): container finished" podID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerID="8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a" exitCode=0 Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.057230 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pv76l" event={"ID":"55ca888a-bbfa-477d-9769-8eab8a6aee98","Type":"ContainerDied","Data":"8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.057251 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pv76l" event={"ID":"55ca888a-bbfa-477d-9769-8eab8a6aee98","Type":"ContainerDied","Data":"8a20183b0f15bee1b9ae76d010ce117a6e0a627fb22f79bdd64e81bdf3dd72a0"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.057293 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pv76l" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.065821 4690 generic.go:334] "Generic (PLEG): container finished" podID="4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" containerID="9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f" exitCode=0 Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.065881 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" event={"ID":"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c","Type":"ContainerDied","Data":"9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.065907 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" event={"ID":"4b79afc2-a76d-4fb3-8be6-ec2d3902b33c","Type":"ContainerDied","Data":"81f2927e009458c5460d2a405d1134cd47d6306812010b635bbb5f903a9ef665"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.065966 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gxcpk" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.069971 4690 generic.go:334] "Generic (PLEG): container finished" podID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerID="5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec" exitCode=0 Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.070006 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcmdk" event={"ID":"4c5d01e4-052c-4541-a385-79a8066bba8f","Type":"ContainerDied","Data":"5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.070025 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcmdk" event={"ID":"4c5d01e4-052c-4541-a385-79a8066bba8f","Type":"ContainerDied","Data":"05e64fbfb40bd1453488500198377bf85b8081227c06b90443ac70705f7e4fc9"} Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.070084 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcmdk" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.077288 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" podStartSLOduration=1.077273499 podStartE2EDuration="1.077273499s" podCreationTimestamp="2026-02-17 00:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:09:16.075043243 +0000 UTC m=+229.810755994" watchObservedRunningTime="2026-02-17 00:09:16.077273499 +0000 UTC m=+229.812986250" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.088536 4690 scope.go:117] "RemoveContainer" containerID="96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.108126 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xntcs"] Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.112743 4690 scope.go:117] "RemoveContainer" containerID="397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.114208 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xntcs"] Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.114888 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b\": container with ID starting with 397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b not found: ID does not exist" containerID="397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.114921 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b"} err="failed to get container status \"397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b\": rpc error: code = NotFound desc = could not find container \"397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b\": container with ID starting with 397c3754640901803f935b719c5508bde622e54a0652f58ddd74cd52cf243a5b not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.114950 4690 scope.go:117] "RemoveContainer" containerID="e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.115423 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82\": container with ID starting with e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82 not found: ID does not exist" containerID="e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.115444 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82"} err="failed to get container status \"e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82\": rpc error: code = NotFound desc = could not find container \"e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82\": container with ID starting with e4ec9690f0b17464a63065347b56d607738820b1db7328025643b8b0c52c2e82 not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.115456 4690 scope.go:117] "RemoveContainer" containerID="96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.115940 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306\": container with ID starting with 96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306 not found: ID does not exist" containerID="96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.115956 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306"} err="failed to get container status \"96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306\": rpc error: code = NotFound desc = could not find container \"96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306\": container with ID starting with 96148c298621d638bdf03c2914e50f1ddba4db962d0473705eaccb30daca4306 not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.115968 4690 scope.go:117] "RemoveContainer" containerID="7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.139157 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2tkvq"] Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.141846 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2tkvq"] Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.150410 4690 scope.go:117] "RemoveContainer" containerID="5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.156429 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pv76l"] Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.160174 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pv76l"] Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.164288 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gxcpk"] Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.166649 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gxcpk"] Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.172166 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fcmdk"] Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.172853 4690 scope.go:117] "RemoveContainer" containerID="97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.175868 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fcmdk"] Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.187014 4690 scope.go:117] "RemoveContainer" containerID="7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.188615 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2\": container with ID starting with 7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2 not found: ID does not exist" containerID="7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.188697 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2"} err="failed to get container status \"7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2\": rpc error: code = NotFound desc = could not find container \"7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2\": container with ID starting with 7df97fdaf8be3cc494252624c4dd50404fa8c5b9ea0f744e0604a5227fb8bbf2 not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.188768 4690 scope.go:117] "RemoveContainer" containerID="5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.189138 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4\": container with ID starting with 5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4 not found: ID does not exist" containerID="5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.189180 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4"} err="failed to get container status \"5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4\": rpc error: code = NotFound desc = could not find container \"5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4\": container with ID starting with 5facdaad3e66bf30040425588ce4a96e1aac17a66e5a315a917c2e00fb1a3cd4 not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.189196 4690 scope.go:117] "RemoveContainer" containerID="97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.189481 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d\": container with ID starting with 97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d not found: ID does not exist" containerID="97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.189502 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d"} err="failed to get container status \"97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d\": rpc error: code = NotFound desc = could not find container \"97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d\": container with ID starting with 97f20fb63a822f313078852cfb9076011b6e7e8fbcba64bbefd82e57839ea57d not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.189517 4690 scope.go:117] "RemoveContainer" containerID="8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.220308 4690 scope.go:117] "RemoveContainer" containerID="e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.237499 4690 scope.go:117] "RemoveContainer" containerID="75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.253438 4690 scope.go:117] "RemoveContainer" containerID="8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.253997 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a\": container with ID starting with 8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a not found: ID does not exist" containerID="8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.254258 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a"} err="failed to get container status \"8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a\": rpc error: code = NotFound desc = could not find container \"8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a\": container with ID starting with 8b7bfcc62cc3d5e0f9d99a641f476ce4f39a411aa6b2e4794c8cb90641819c1a not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.254297 4690 scope.go:117] "RemoveContainer" containerID="e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.254682 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf\": container with ID starting with e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf not found: ID does not exist" containerID="e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.254714 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf"} err="failed to get container status \"e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf\": rpc error: code = NotFound desc = could not find container \"e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf\": container with ID starting with e2f585492e9e132992d5fab3c5f87dbabc9c6be1679f0f1e118e03756ede7fdf not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.254737 4690 scope.go:117] "RemoveContainer" containerID="75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.255035 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a\": container with ID starting with 75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a not found: ID does not exist" containerID="75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.255061 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a"} err="failed to get container status \"75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a\": rpc error: code = NotFound desc = could not find container \"75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a\": container with ID starting with 75580d511438fbdc23d1c12dd96e8650f9d7786e013a74ab6a414a045020806a not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.255074 4690 scope.go:117] "RemoveContainer" containerID="9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.271214 4690 scope.go:117] "RemoveContainer" containerID="9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.271639 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f\": container with ID starting with 9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f not found: ID does not exist" containerID="9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.271685 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f"} err="failed to get container status \"9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f\": rpc error: code = NotFound desc = could not find container \"9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f\": container with ID starting with 9148836b3bcd58db63017348174b978ca5771ffd19ee70b215f698f21c6c261f not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.271717 4690 scope.go:117] "RemoveContainer" containerID="5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.285573 4690 scope.go:117] "RemoveContainer" containerID="eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.303470 4690 scope.go:117] "RemoveContainer" containerID="b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.318641 4690 scope.go:117] "RemoveContainer" containerID="5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.319492 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec\": container with ID starting with 5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec not found: ID does not exist" containerID="5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.319531 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec"} err="failed to get container status \"5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec\": rpc error: code = NotFound desc = could not find container \"5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec\": container with ID starting with 5a746818ef9b2aeb09b7b6f889a7e6fea0b031e2ddcb77c19216d5a5b478faec not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.319564 4690 scope.go:117] "RemoveContainer" containerID="eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.319912 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459\": container with ID starting with eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459 not found: ID does not exist" containerID="eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.319950 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459"} err="failed to get container status \"eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459\": rpc error: code = NotFound desc = could not find container \"eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459\": container with ID starting with eeb3e11bdfafc8475c3fd5a108b502934c8743f0c36ca92e673d0eef7e712459 not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.319992 4690 scope.go:117] "RemoveContainer" containerID="b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.320393 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07\": container with ID starting with b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07 not found: ID does not exist" containerID="b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.320416 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07"} err="failed to get container status \"b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07\": rpc error: code = NotFound desc = could not find container \"b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07\": container with ID starting with b4f31e96624d12087eff6f4aa7ec6ec944d7dad04ae1c61dcc355587e3871f07 not found: ID does not exist" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.797871 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lqg9n"] Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798125 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerName="extract-utilities" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798141 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerName="extract-utilities" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798152 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08acd43-80f4-4a48-b848-7899561e26b5" containerName="extract-utilities" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798160 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08acd43-80f4-4a48-b848-7899561e26b5" containerName="extract-utilities" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798171 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerName="extract-content" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798180 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerName="extract-content" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798190 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerName="extract-content" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798197 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerName="extract-content" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798236 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5df6322-f224-4e0e-8099-02865f10660a" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798247 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5df6322-f224-4e0e-8099-02865f10660a" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798257 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798264 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798275 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5df6322-f224-4e0e-8099-02865f10660a" containerName="extract-utilities" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798282 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5df6322-f224-4e0e-8099-02865f10660a" containerName="extract-utilities" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798292 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerName="extract-utilities" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798298 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerName="extract-utilities" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798309 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08acd43-80f4-4a48-b848-7899561e26b5" containerName="extract-content" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798318 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08acd43-80f4-4a48-b848-7899561e26b5" containerName="extract-content" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798330 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5df6322-f224-4e0e-8099-02865f10660a" containerName="extract-content" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798338 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5df6322-f224-4e0e-8099-02865f10660a" containerName="extract-content" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798350 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798375 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798386 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" containerName="marketplace-operator" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798393 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" containerName="marketplace-operator" Feb 17 00:09:16 crc kubenswrapper[4690]: E0217 00:09:16.798403 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08acd43-80f4-4a48-b848-7899561e26b5" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798412 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08acd43-80f4-4a48-b848-7899561e26b5" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798508 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5df6322-f224-4e0e-8099-02865f10660a" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798523 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" containerName="marketplace-operator" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798532 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5d01e4-052c-4541-a385-79a8066bba8f" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798543 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="55ca888a-bbfa-477d-9769-8eab8a6aee98" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.798554 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c08acd43-80f4-4a48-b848-7899561e26b5" containerName="registry-server" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.799422 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.803260 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.813302 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lqg9n"] Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.926790 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfkg5\" (UniqueName: \"kubernetes.io/projected/9debb4ff-2d87-49ea-a168-8b403ef878cf-kube-api-access-sfkg5\") pod \"community-operators-lqg9n\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.926852 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-catalog-content\") pod \"community-operators-lqg9n\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.926913 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-utilities\") pod \"community-operators-lqg9n\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.990656 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b79afc2-a76d-4fb3-8be6-ec2d3902b33c" path="/var/lib/kubelet/pods/4b79afc2-a76d-4fb3-8be6-ec2d3902b33c/volumes" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.991404 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c5d01e4-052c-4541-a385-79a8066bba8f" path="/var/lib/kubelet/pods/4c5d01e4-052c-4541-a385-79a8066bba8f/volumes" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.992289 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55ca888a-bbfa-477d-9769-8eab8a6aee98" path="/var/lib/kubelet/pods/55ca888a-bbfa-477d-9769-8eab8a6aee98/volumes" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.993819 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c08acd43-80f4-4a48-b848-7899561e26b5" path="/var/lib/kubelet/pods/c08acd43-80f4-4a48-b848-7899561e26b5/volumes" Feb 17 00:09:16 crc kubenswrapper[4690]: I0217 00:09:16.995652 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5df6322-f224-4e0e-8099-02865f10660a" path="/var/lib/kubelet/pods/e5df6322-f224-4e0e-8099-02865f10660a/volumes" Feb 17 00:09:17 crc kubenswrapper[4690]: I0217 00:09:17.028906 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfkg5\" (UniqueName: \"kubernetes.io/projected/9debb4ff-2d87-49ea-a168-8b403ef878cf-kube-api-access-sfkg5\") pod \"community-operators-lqg9n\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:17 crc kubenswrapper[4690]: I0217 00:09:17.028960 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-catalog-content\") pod \"community-operators-lqg9n\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:17 crc kubenswrapper[4690]: I0217 00:09:17.029032 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-utilities\") pod \"community-operators-lqg9n\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:17 crc kubenswrapper[4690]: I0217 00:09:17.029831 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-catalog-content\") pod \"community-operators-lqg9n\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:17 crc kubenswrapper[4690]: I0217 00:09:17.029992 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-utilities\") pod \"community-operators-lqg9n\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:17 crc kubenswrapper[4690]: I0217 00:09:17.055723 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfkg5\" (UniqueName: \"kubernetes.io/projected/9debb4ff-2d87-49ea-a168-8b403ef878cf-kube-api-access-sfkg5\") pod \"community-operators-lqg9n\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:17 crc kubenswrapper[4690]: I0217 00:09:17.100234 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wvvgs" Feb 17 00:09:17 crc kubenswrapper[4690]: I0217 00:09:17.119211 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:17 crc kubenswrapper[4690]: I0217 00:09:17.520672 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lqg9n"] Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.103106 4690 generic.go:334] "Generic (PLEG): container finished" podID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerID="6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6" exitCode=0 Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.105695 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lqg9n" event={"ID":"9debb4ff-2d87-49ea-a168-8b403ef878cf","Type":"ContainerDied","Data":"6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6"} Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.105851 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lqg9n" event={"ID":"9debb4ff-2d87-49ea-a168-8b403ef878cf","Type":"ContainerStarted","Data":"05b831284693454541df36a8e58a28d8f0b1be6baaf046688faaa7fd1e63d82c"} Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.197329 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fqtk7"] Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.198651 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.202592 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.205278 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqtk7"] Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.255214 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-catalog-content\") pod \"redhat-marketplace-fqtk7\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.255271 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-utilities\") pod \"redhat-marketplace-fqtk7\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.255296 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmdz6\" (UniqueName: \"kubernetes.io/projected/8151c43f-c4a2-4b6e-856d-298a43bd2f17-kube-api-access-nmdz6\") pod \"redhat-marketplace-fqtk7\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.356911 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-catalog-content\") pod \"redhat-marketplace-fqtk7\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.356988 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-utilities\") pod \"redhat-marketplace-fqtk7\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.357020 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmdz6\" (UniqueName: \"kubernetes.io/projected/8151c43f-c4a2-4b6e-856d-298a43bd2f17-kube-api-access-nmdz6\") pod \"redhat-marketplace-fqtk7\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.358151 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-catalog-content\") pod \"redhat-marketplace-fqtk7\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.358203 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-utilities\") pod \"redhat-marketplace-fqtk7\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.377613 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmdz6\" (UniqueName: \"kubernetes.io/projected/8151c43f-c4a2-4b6e-856d-298a43bd2f17-kube-api-access-nmdz6\") pod \"redhat-marketplace-fqtk7\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.523666 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:18 crc kubenswrapper[4690]: I0217 00:09:18.995068 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqtk7"] Feb 17 00:09:19 crc kubenswrapper[4690]: W0217 00:09:19.003933 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8151c43f_c4a2_4b6e_856d_298a43bd2f17.slice/crio-edc2658e000e8940e7072beae0ee0ccc2669c7d341ab1b7fe4fd862015102064 WatchSource:0}: Error finding container edc2658e000e8940e7072beae0ee0ccc2669c7d341ab1b7fe4fd862015102064: Status 404 returned error can't find the container with id edc2658e000e8940e7072beae0ee0ccc2669c7d341ab1b7fe4fd862015102064 Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.111331 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqtk7" event={"ID":"8151c43f-c4a2-4b6e-856d-298a43bd2f17","Type":"ContainerStarted","Data":"edc2658e000e8940e7072beae0ee0ccc2669c7d341ab1b7fe4fd862015102064"} Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.113349 4690 generic.go:334] "Generic (PLEG): container finished" podID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerID="780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51" exitCode=0 Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.113453 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lqg9n" event={"ID":"9debb4ff-2d87-49ea-a168-8b403ef878cf","Type":"ContainerDied","Data":"780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51"} Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.199992 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qnd7w"] Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.201799 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.203237 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qnd7w"] Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.203801 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.269593 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cb4a9ed-9d67-4635-9083-9c6e383f148f-utilities\") pod \"redhat-operators-qnd7w\" (UID: \"2cb4a9ed-9d67-4635-9083-9c6e383f148f\") " pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.269643 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cb4a9ed-9d67-4635-9083-9c6e383f148f-catalog-content\") pod \"redhat-operators-qnd7w\" (UID: \"2cb4a9ed-9d67-4635-9083-9c6e383f148f\") " pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.269673 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw886\" (UniqueName: \"kubernetes.io/projected/2cb4a9ed-9d67-4635-9083-9c6e383f148f-kube-api-access-cw886\") pod \"redhat-operators-qnd7w\" (UID: \"2cb4a9ed-9d67-4635-9083-9c6e383f148f\") " pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.370242 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw886\" (UniqueName: \"kubernetes.io/projected/2cb4a9ed-9d67-4635-9083-9c6e383f148f-kube-api-access-cw886\") pod \"redhat-operators-qnd7w\" (UID: \"2cb4a9ed-9d67-4635-9083-9c6e383f148f\") " pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.370372 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cb4a9ed-9d67-4635-9083-9c6e383f148f-utilities\") pod \"redhat-operators-qnd7w\" (UID: \"2cb4a9ed-9d67-4635-9083-9c6e383f148f\") " pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.370432 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cb4a9ed-9d67-4635-9083-9c6e383f148f-catalog-content\") pod \"redhat-operators-qnd7w\" (UID: \"2cb4a9ed-9d67-4635-9083-9c6e383f148f\") " pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.370917 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cb4a9ed-9d67-4635-9083-9c6e383f148f-catalog-content\") pod \"redhat-operators-qnd7w\" (UID: \"2cb4a9ed-9d67-4635-9083-9c6e383f148f\") " pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.371198 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cb4a9ed-9d67-4635-9083-9c6e383f148f-utilities\") pod \"redhat-operators-qnd7w\" (UID: \"2cb4a9ed-9d67-4635-9083-9c6e383f148f\") " pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.391614 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw886\" (UniqueName: \"kubernetes.io/projected/2cb4a9ed-9d67-4635-9083-9c6e383f148f-kube-api-access-cw886\") pod \"redhat-operators-qnd7w\" (UID: \"2cb4a9ed-9d67-4635-9083-9c6e383f148f\") " pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.591734 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:19 crc kubenswrapper[4690]: I0217 00:09:19.788391 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qnd7w"] Feb 17 00:09:19 crc kubenswrapper[4690]: W0217 00:09:19.796679 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cb4a9ed_9d67_4635_9083_9c6e383f148f.slice/crio-cac103750f7cb46a537b541175444ab4bbc8749154d9d9f7ba3b1503d231e1f1 WatchSource:0}: Error finding container cac103750f7cb46a537b541175444ab4bbc8749154d9d9f7ba3b1503d231e1f1: Status 404 returned error can't find the container with id cac103750f7cb46a537b541175444ab4bbc8749154d9d9f7ba3b1503d231e1f1 Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.123610 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lqg9n" event={"ID":"9debb4ff-2d87-49ea-a168-8b403ef878cf","Type":"ContainerStarted","Data":"2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd"} Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.126529 4690 generic.go:334] "Generic (PLEG): container finished" podID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerID="05a35ed956445475b96f91928f11420e1eda12c4f568a66eaf4ee29a73512741" exitCode=0 Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.126613 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqtk7" event={"ID":"8151c43f-c4a2-4b6e-856d-298a43bd2f17","Type":"ContainerDied","Data":"05a35ed956445475b96f91928f11420e1eda12c4f568a66eaf4ee29a73512741"} Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.130736 4690 generic.go:334] "Generic (PLEG): container finished" podID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" containerID="74fd0cf363fe41a2333b49fc960bddcdfed0b0a73f383ebc650549f754335046" exitCode=0 Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.130808 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnd7w" event={"ID":"2cb4a9ed-9d67-4635-9083-9c6e383f148f","Type":"ContainerDied","Data":"74fd0cf363fe41a2333b49fc960bddcdfed0b0a73f383ebc650549f754335046"} Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.130843 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnd7w" event={"ID":"2cb4a9ed-9d67-4635-9083-9c6e383f148f","Type":"ContainerStarted","Data":"cac103750f7cb46a537b541175444ab4bbc8749154d9d9f7ba3b1503d231e1f1"} Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.162032 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lqg9n" podStartSLOduration=2.767049259 podStartE2EDuration="4.162004972s" podCreationTimestamp="2026-02-17 00:09:16 +0000 UTC" firstStartedPulling="2026-02-17 00:09:18.106376637 +0000 UTC m=+231.842089388" lastFinishedPulling="2026-02-17 00:09:19.50133234 +0000 UTC m=+233.237045101" observedRunningTime="2026-02-17 00:09:20.145759981 +0000 UTC m=+233.881472732" watchObservedRunningTime="2026-02-17 00:09:20.162004972 +0000 UTC m=+233.897717723" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.593184 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4glwj"] Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.594442 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.596186 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.606880 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4glwj"] Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.690685 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b18214-ded8-4377-81de-0c21887f2576-catalog-content\") pod \"certified-operators-4glwj\" (UID: \"c8b18214-ded8-4377-81de-0c21887f2576\") " pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.690739 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b18214-ded8-4377-81de-0c21887f2576-utilities\") pod \"certified-operators-4glwj\" (UID: \"c8b18214-ded8-4377-81de-0c21887f2576\") " pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.690776 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf8pq\" (UniqueName: \"kubernetes.io/projected/c8b18214-ded8-4377-81de-0c21887f2576-kube-api-access-rf8pq\") pod \"certified-operators-4glwj\" (UID: \"c8b18214-ded8-4377-81de-0c21887f2576\") " pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.792285 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b18214-ded8-4377-81de-0c21887f2576-catalog-content\") pod \"certified-operators-4glwj\" (UID: \"c8b18214-ded8-4377-81de-0c21887f2576\") " pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.792328 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b18214-ded8-4377-81de-0c21887f2576-utilities\") pod \"certified-operators-4glwj\" (UID: \"c8b18214-ded8-4377-81de-0c21887f2576\") " pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.792373 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf8pq\" (UniqueName: \"kubernetes.io/projected/c8b18214-ded8-4377-81de-0c21887f2576-kube-api-access-rf8pq\") pod \"certified-operators-4glwj\" (UID: \"c8b18214-ded8-4377-81de-0c21887f2576\") " pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.793017 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b18214-ded8-4377-81de-0c21887f2576-catalog-content\") pod \"certified-operators-4glwj\" (UID: \"c8b18214-ded8-4377-81de-0c21887f2576\") " pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.793192 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b18214-ded8-4377-81de-0c21887f2576-utilities\") pod \"certified-operators-4glwj\" (UID: \"c8b18214-ded8-4377-81de-0c21887f2576\") " pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.811514 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf8pq\" (UniqueName: \"kubernetes.io/projected/c8b18214-ded8-4377-81de-0c21887f2576-kube-api-access-rf8pq\") pod \"certified-operators-4glwj\" (UID: \"c8b18214-ded8-4377-81de-0c21887f2576\") " pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:20 crc kubenswrapper[4690]: I0217 00:09:20.917015 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:21 crc kubenswrapper[4690]: I0217 00:09:21.139864 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4glwj"] Feb 17 00:09:21 crc kubenswrapper[4690]: I0217 00:09:21.144149 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnd7w" event={"ID":"2cb4a9ed-9d67-4635-9083-9c6e383f148f","Type":"ContainerStarted","Data":"a00c5d1f9d4965aa7b75e39d383a732e64dacea0bedafbf3504b17985b070775"} Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.155551 4690 generic.go:334] "Generic (PLEG): container finished" podID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerID="94c9dd6933eef1de8e551b4403f67043b9607d8b45bf9fcb6519630a355b40b4" exitCode=0 Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.155632 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqtk7" event={"ID":"8151c43f-c4a2-4b6e-856d-298a43bd2f17","Type":"ContainerDied","Data":"94c9dd6933eef1de8e551b4403f67043b9607d8b45bf9fcb6519630a355b40b4"} Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.161947 4690 generic.go:334] "Generic (PLEG): container finished" podID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" containerID="a00c5d1f9d4965aa7b75e39d383a732e64dacea0bedafbf3504b17985b070775" exitCode=0 Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.162014 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnd7w" event={"ID":"2cb4a9ed-9d67-4635-9083-9c6e383f148f","Type":"ContainerDied","Data":"a00c5d1f9d4965aa7b75e39d383a732e64dacea0bedafbf3504b17985b070775"} Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.163787 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8b18214-ded8-4377-81de-0c21887f2576" containerID="a18185cdcea3eb3432b23926caa72cf7ee889abf1cce49c6d3d142e287412969" exitCode=0 Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.163821 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4glwj" event={"ID":"c8b18214-ded8-4377-81de-0c21887f2576","Type":"ContainerDied","Data":"a18185cdcea3eb3432b23926caa72cf7ee889abf1cce49c6d3d142e287412969"} Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.163844 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4glwj" event={"ID":"c8b18214-ded8-4377-81de-0c21887f2576","Type":"ContainerStarted","Data":"466e4fe2e356e4e6b4eb94cd1f8d8f0c158814110dd45b39034172ebf02729d4"} Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.339932 4690 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343396 4690 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343457 4690 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 17 00:09:22 crc kubenswrapper[4690]: E0217 00:09:22.343650 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343678 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 00:09:22 crc kubenswrapper[4690]: E0217 00:09:22.343689 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343699 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 17 00:09:22 crc kubenswrapper[4690]: E0217 00:09:22.343720 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343729 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 17 00:09:22 crc kubenswrapper[4690]: E0217 00:09:22.343740 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343749 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 17 00:09:22 crc kubenswrapper[4690]: E0217 00:09:22.343762 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343770 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 17 00:09:22 crc kubenswrapper[4690]: E0217 00:09:22.343781 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343789 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 00:09:22 crc kubenswrapper[4690]: E0217 00:09:22.343803 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343811 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343927 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343942 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343951 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343961 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.343974 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.344220 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.345760 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.347147 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d" gracePeriod=15 Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.347405 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b" gracePeriod=15 Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.347404 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3" gracePeriod=15 Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.349784 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef" gracePeriod=15 Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.347611 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d" gracePeriod=15 Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.351533 4690 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.421079 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.421137 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.421165 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.421267 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.421421 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.421519 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.421640 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.421800 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.523393 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.523787 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.523834 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.523864 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.523892 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.523913 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.523937 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.523977 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.524069 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.523557 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.524118 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.524141 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.524162 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.524181 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.524209 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: I0217 00:09:22.524236 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:22 crc kubenswrapper[4690]: E0217 00:09:22.528456 4690 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.181:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-fqtk7.1894e01b94bf22f6 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-fqtk7,UID:8151c43f-c4a2-4b6e-856d-298a43bd2f17,APIVersion:v1,ResourceVersion:29681,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 370ms (370ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-17 00:09:22.527740662 +0000 UTC m=+236.263453413,LastTimestamp:2026-02-17 00:09:22.527740662 +0000 UTC m=+236.263453413,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 17 00:09:22 crc kubenswrapper[4690]: E0217 00:09:22.736483 4690 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.181:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-fqtk7.1894e01b94bf22f6 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-fqtk7,UID:8151c43f-c4a2-4b6e-856d-298a43bd2f17,APIVersion:v1,ResourceVersion:29681,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 370ms (370ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-17 00:09:22.527740662 +0000 UTC m=+236.263453413,LastTimestamp:2026-02-17 00:09:22.527740662 +0000 UTC m=+236.263453413,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.169630 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4glwj" event={"ID":"c8b18214-ded8-4377-81de-0c21887f2576","Type":"ContainerStarted","Data":"de637bdb2ed27db14f6497872e2027829005b4856a3b8ce176067f53464ee47f"} Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.170590 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.171447 4690 generic.go:334] "Generic (PLEG): container finished" podID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" containerID="f0d770eb77c3a1d0c44dc40b6327a87813001b1e1c5d84d96eaa1d544d3eaa9e" exitCode=0 Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.171526 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d36eab40-817c-4939-a0d6-ea1fbb2949ca","Type":"ContainerDied","Data":"f0d770eb77c3a1d0c44dc40b6327a87813001b1e1c5d84d96eaa1d544d3eaa9e"} Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.171969 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.172223 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.173100 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqtk7" event={"ID":"8151c43f-c4a2-4b6e-856d-298a43bd2f17","Type":"ContainerStarted","Data":"9653deb103ebbb851ce804a8febe6d315cf14b53883caf69688abf4682840234"} Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.173476 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.173694 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.174008 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.174971 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.176030 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.176644 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef" exitCode=0 Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.176664 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b" exitCode=0 Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.176672 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d" exitCode=0 Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.176682 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3" exitCode=2 Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.176750 4690 scope.go:117] "RemoveContainer" containerID="a74c35fd3c83ba11cb8c04ef5378d0b524c6b476eb36e35f6db3f5a42a2da764" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.181721 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnd7w" event={"ID":"2cb4a9ed-9d67-4635-9083-9c6e383f148f","Type":"ContainerStarted","Data":"728bd900f77bb129156fa4f2285844bd82dd827145ca8229e85c207893f67a96"} Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.182400 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.182667 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.183440 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:23 crc kubenswrapper[4690]: I0217 00:09:23.183896 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:24 crc kubenswrapper[4690]: I0217 00:09:24.193606 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 17 00:09:24 crc kubenswrapper[4690]: I0217 00:09:24.196547 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8b18214-ded8-4377-81de-0c21887f2576" containerID="de637bdb2ed27db14f6497872e2027829005b4856a3b8ce176067f53464ee47f" exitCode=0 Feb 17 00:09:24 crc kubenswrapper[4690]: I0217 00:09:24.196992 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4glwj" event={"ID":"c8b18214-ded8-4377-81de-0c21887f2576","Type":"ContainerDied","Data":"de637bdb2ed27db14f6497872e2027829005b4856a3b8ce176067f53464ee47f"} Feb 17 00:09:24 crc kubenswrapper[4690]: I0217 00:09:24.197817 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:24 crc kubenswrapper[4690]: I0217 00:09:24.198235 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:24 crc kubenswrapper[4690]: I0217 00:09:24.199099 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:24 crc kubenswrapper[4690]: I0217 00:09:24.199596 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.203162 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d36eab40-817c-4939-a0d6-ea1fbb2949ca","Type":"ContainerDied","Data":"25beef08c99c361d59a6124afdb6d666cb1f6220d06dc7ad90f71d69857334ba"} Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.203962 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25beef08c99c361d59a6124afdb6d666cb1f6220d06dc7ad90f71d69857334ba" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.207350 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.208273 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d" exitCode=0 Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.208326 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c64f5af11ef65dafb23db09ad46cbc5ff9545557b723769bc6ec3d036835b9bd" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.516564 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.517268 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.517506 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.517780 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.518010 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.522625 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.523341 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.523729 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.524093 4690 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.524294 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.524523 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.524824 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560005 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560094 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-var-lock\") pod \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560127 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560156 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kubelet-dir\") pod \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560171 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-var-lock" (OuterVolumeSpecName: "var-lock") pod "d36eab40-817c-4939-a0d6-ea1fbb2949ca" (UID: "d36eab40-817c-4939-a0d6-ea1fbb2949ca"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560213 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560241 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560266 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d36eab40-817c-4939-a0d6-ea1fbb2949ca" (UID: "d36eab40-817c-4939-a0d6-ea1fbb2949ca"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560279 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kube-api-access\") pod \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\" (UID: \"d36eab40-817c-4939-a0d6-ea1fbb2949ca\") " Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560292 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560310 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560522 4690 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560539 4690 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560550 4690 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560560 4690 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-var-lock\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.560569 4690 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.565961 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d36eab40-817c-4939-a0d6-ea1fbb2949ca" (UID: "d36eab40-817c-4939-a0d6-ea1fbb2949ca"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:09:25 crc kubenswrapper[4690]: I0217 00:09:25.661382 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d36eab40-817c-4939-a0d6-ea1fbb2949ca-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.216051 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4glwj" event={"ID":"c8b18214-ded8-4377-81de-0c21887f2576","Type":"ContainerStarted","Data":"e73510c7870d96afa35c4730da2c5453ba780b1b40165530f293560ab4cc909e"} Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.216140 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.216178 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.216815 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.217249 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.217815 4690 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.218152 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.218549 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.235792 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.236274 4690 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.236618 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.237014 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.237221 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.237482 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.237784 4690 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.238025 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.238267 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.238564 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.978556 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.980204 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.980501 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.986695 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.987436 4690 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:26 crc kubenswrapper[4690]: I0217 00:09:26.987665 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.121321 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.122599 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.170508 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.171449 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.171764 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.172196 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.172683 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.173218 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.263822 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.264445 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.265878 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.266185 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.266695 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.267030 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: E0217 00:09:27.337255 4690 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: E0217 00:09:27.337743 4690 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: E0217 00:09:27.338245 4690 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: E0217 00:09:27.338576 4690 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: E0217 00:09:27.341550 4690 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.341593 4690 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 17 00:09:27 crc kubenswrapper[4690]: E0217 00:09:27.341900 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="200ms" Feb 17 00:09:27 crc kubenswrapper[4690]: E0217 00:09:27.397925 4690 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.181:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:27 crc kubenswrapper[4690]: I0217 00:09:27.398343 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:27 crc kubenswrapper[4690]: W0217 00:09:27.421195 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-ecf391a13a5c0ec5522d4ea78dbb71f95070f61d9474ee6b2489ad8032479f51 WatchSource:0}: Error finding container ecf391a13a5c0ec5522d4ea78dbb71f95070f61d9474ee6b2489ad8032479f51: Status 404 returned error can't find the container with id ecf391a13a5c0ec5522d4ea78dbb71f95070f61d9474ee6b2489ad8032479f51 Feb 17 00:09:27 crc kubenswrapper[4690]: E0217 00:09:27.543073 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="400ms" Feb 17 00:09:27 crc kubenswrapper[4690]: E0217 00:09:27.945036 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="800ms" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.237685 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"32fe2039e895ae61d9818f5ae9926b9200de3bf091b06e767fc5568b9e3148d0"} Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.237740 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ecf391a13a5c0ec5522d4ea78dbb71f95070f61d9474ee6b2489ad8032479f51"} Feb 17 00:09:28 crc kubenswrapper[4690]: E0217 00:09:28.238801 4690 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.181:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.239238 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.239591 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.240025 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.243475 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.243902 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: E0217 00:09:28.383503 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:09:28Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:09:28Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:09:28Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T00:09:28Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: E0217 00:09:28.384146 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: E0217 00:09:28.385098 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: E0217 00:09:28.385581 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: E0217 00:09:28.385931 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: E0217 00:09:28.385969 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.523959 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.524007 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.576067 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.576448 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.576647 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.577476 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.577863 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: I0217 00:09:28.578254 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:28 crc kubenswrapper[4690]: E0217 00:09:28.746506 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="1.6s" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.299014 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.299622 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.299967 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.300474 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.300698 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.300954 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.592925 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.594409 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.661484 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.661948 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.662233 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.662692 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.663115 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:29 crc kubenswrapper[4690]: I0217 00:09:29.663398 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.304407 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qnd7w" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.305111 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.305562 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.305843 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.306896 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.307095 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:30 crc kubenswrapper[4690]: E0217 00:09:30.348067 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="3.2s" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.918175 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.918892 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.961862 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.962279 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.962586 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.962819 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.963025 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:30 crc kubenswrapper[4690]: I0217 00:09:30.963240 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:31 crc kubenswrapper[4690]: I0217 00:09:31.298958 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4glwj" Feb 17 00:09:31 crc kubenswrapper[4690]: I0217 00:09:31.299504 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:31 crc kubenswrapper[4690]: I0217 00:09:31.299982 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:31 crc kubenswrapper[4690]: I0217 00:09:31.300401 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:31 crc kubenswrapper[4690]: I0217 00:09:31.300682 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:31 crc kubenswrapper[4690]: I0217 00:09:31.300986 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:32 crc kubenswrapper[4690]: E0217 00:09:32.737681 4690 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.181:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-fqtk7.1894e01b94bf22f6 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-fqtk7,UID:8151c43f-c4a2-4b6e-856d-298a43bd2f17,APIVersion:v1,ResourceVersion:29681,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 370ms (370ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-17 00:09:22.527740662 +0000 UTC m=+236.263453413,LastTimestamp:2026-02-17 00:09:22.527740662 +0000 UTC m=+236.263453413,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 17 00:09:33 crc kubenswrapper[4690]: E0217 00:09:33.549785 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="6.4s" Feb 17 00:09:34 crc kubenswrapper[4690]: I0217 00:09:34.975853 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:34 crc kubenswrapper[4690]: I0217 00:09:34.977072 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:34 crc kubenswrapper[4690]: I0217 00:09:34.977545 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:34 crc kubenswrapper[4690]: I0217 00:09:34.977955 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:34 crc kubenswrapper[4690]: I0217 00:09:34.978247 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:34 crc kubenswrapper[4690]: I0217 00:09:34.978491 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:35 crc kubenswrapper[4690]: I0217 00:09:35.002952 4690 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac3fd592-6a9c-44aa-87bd-31732029af22" Feb 17 00:09:35 crc kubenswrapper[4690]: I0217 00:09:35.003001 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac3fd592-6a9c-44aa-87bd-31732029af22" Feb 17 00:09:35 crc kubenswrapper[4690]: E0217 00:09:35.003567 4690 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:35 crc kubenswrapper[4690]: I0217 00:09:35.004321 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:35 crc kubenswrapper[4690]: W0217 00:09:35.039833 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-8c7f5fc306b1b054a2f494034a721deb1ce6ffc26df02afc5ca7588be21de691 WatchSource:0}: Error finding container 8c7f5fc306b1b054a2f494034a721deb1ce6ffc26df02afc5ca7588be21de691: Status 404 returned error can't find the container with id 8c7f5fc306b1b054a2f494034a721deb1ce6ffc26df02afc5ca7588be21de691 Feb 17 00:09:35 crc kubenswrapper[4690]: I0217 00:09:35.282600 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8c7f5fc306b1b054a2f494034a721deb1ce6ffc26df02afc5ca7588be21de691"} Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.290259 4690 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="c45e653700f959fae0c623861560f7290262d4b7c5444f8c25a183151f9cb119" exitCode=0 Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.290319 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"c45e653700f959fae0c623861560f7290262d4b7c5444f8c25a183151f9cb119"} Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.290721 4690 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac3fd592-6a9c-44aa-87bd-31732029af22" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.290750 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac3fd592-6a9c-44aa-87bd-31732029af22" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.291098 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:36 crc kubenswrapper[4690]: E0217 00:09:36.291324 4690 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.291338 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.291703 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.292066 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.292389 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.293292 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.293332 4690 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2" exitCode=1 Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.293368 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2"} Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.293817 4690 scope.go:117] "RemoveContainer" containerID="c66d9662784df5bca8743983018f6cc0af0a60dc118c7c598d25837dee9591f2" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.294100 4690 status_manager.go:851] "Failed to get status for pod" podUID="2cb4a9ed-9d67-4635-9083-9c6e383f148f" pod="openshift-marketplace/redhat-operators-qnd7w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qnd7w\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.294535 4690 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.294790 4690 status_manager.go:851] "Failed to get status for pod" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" pod="openshift-marketplace/community-operators-lqg9n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-lqg9n\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.295078 4690 status_manager.go:851] "Failed to get status for pod" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" pod="openshift-marketplace/redhat-marketplace-fqtk7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fqtk7\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.295391 4690 status_manager.go:851] "Failed to get status for pod" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:36 crc kubenswrapper[4690]: I0217 00:09:36.295638 4690 status_manager.go:851] "Failed to get status for pod" podUID="c8b18214-ded8-4377-81de-0c21887f2576" pod="openshift-marketplace/certified-operators-4glwj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4glwj\": dial tcp 38.102.83.181:6443: connect: connection refused" Feb 17 00:09:37 crc kubenswrapper[4690]: I0217 00:09:37.302561 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4299243a0a91c1192b89b7da2c492d558014c1b24666a85822c036f805b34900"} Feb 17 00:09:37 crc kubenswrapper[4690]: I0217 00:09:37.302853 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6cd2001a130b2accfe6ea9ecdd4e021cf51e34467a85ff8d9ed3afdd4cdbc3dc"} Feb 17 00:09:37 crc kubenswrapper[4690]: I0217 00:09:37.302864 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"090f7b947516af7c09dc30b453a243373e1e275ad51ad2be826fe93ebb8485f4"} Feb 17 00:09:37 crc kubenswrapper[4690]: I0217 00:09:37.302873 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d951b7aa9142d80684b7efa658e3c42dfb271fdd8a73fb5f885b2b29e614630f"} Feb 17 00:09:37 crc kubenswrapper[4690]: I0217 00:09:37.305490 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 17 00:09:37 crc kubenswrapper[4690]: I0217 00:09:37.305547 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"48483891954ade49f00caf9b36914dbe14fb1cb0ae59bd3bb2408cbc94896714"} Feb 17 00:09:38 crc kubenswrapper[4690]: I0217 00:09:38.316683 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1c475abea441fa1acddb9f93fa9ef7b763d5d89264baa9e15fc31d3289bbc90e"} Feb 17 00:09:38 crc kubenswrapper[4690]: I0217 00:09:38.316904 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:38 crc kubenswrapper[4690]: I0217 00:09:38.316999 4690 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac3fd592-6a9c-44aa-87bd-31732029af22" Feb 17 00:09:38 crc kubenswrapper[4690]: I0217 00:09:38.317028 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac3fd592-6a9c-44aa-87bd-31732029af22" Feb 17 00:09:38 crc kubenswrapper[4690]: I0217 00:09:38.600957 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:09:40 crc kubenswrapper[4690]: I0217 00:09:40.004837 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:40 crc kubenswrapper[4690]: I0217 00:09:40.006002 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:40 crc kubenswrapper[4690]: I0217 00:09:40.014268 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:41 crc kubenswrapper[4690]: I0217 00:09:41.713558 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:09:41 crc kubenswrapper[4690]: I0217 00:09:41.713963 4690 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 17 00:09:41 crc kubenswrapper[4690]: I0217 00:09:41.714033 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 17 00:09:43 crc kubenswrapper[4690]: I0217 00:09:43.327714 4690 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:43 crc kubenswrapper[4690]: I0217 00:09:43.344714 4690 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac3fd592-6a9c-44aa-87bd-31732029af22" Feb 17 00:09:43 crc kubenswrapper[4690]: I0217 00:09:43.344747 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac3fd592-6a9c-44aa-87bd-31732029af22" Feb 17 00:09:43 crc kubenswrapper[4690]: I0217 00:09:43.352004 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:43 crc kubenswrapper[4690]: I0217 00:09:43.355102 4690 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="49af898a-c613-4238-9405-268c8cbcf96d" Feb 17 00:09:44 crc kubenswrapper[4690]: I0217 00:09:44.353102 4690 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac3fd592-6a9c-44aa-87bd-31732029af22" Feb 17 00:09:44 crc kubenswrapper[4690]: I0217 00:09:44.353571 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac3fd592-6a9c-44aa-87bd-31732029af22" Feb 17 00:09:47 crc kubenswrapper[4690]: I0217 00:09:47.020249 4690 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="49af898a-c613-4238-9405-268c8cbcf96d" Feb 17 00:09:51 crc kubenswrapper[4690]: I0217 00:09:51.713429 4690 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 17 00:09:51 crc kubenswrapper[4690]: I0217 00:09:51.714132 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 17 00:09:52 crc kubenswrapper[4690]: I0217 00:09:52.471056 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 17 00:09:53 crc kubenswrapper[4690]: I0217 00:09:53.406133 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 17 00:09:53 crc kubenswrapper[4690]: I0217 00:09:53.473796 4690 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 17 00:09:53 crc kubenswrapper[4690]: I0217 00:09:53.478629 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fqtk7" podStartSLOduration=33.079161848 podStartE2EDuration="35.478587576s" podCreationTimestamp="2026-02-17 00:09:18 +0000 UTC" firstStartedPulling="2026-02-17 00:09:20.128306684 +0000 UTC m=+233.864019435" lastFinishedPulling="2026-02-17 00:09:22.527732412 +0000 UTC m=+236.263445163" observedRunningTime="2026-02-17 00:09:43.107718562 +0000 UTC m=+256.843431323" watchObservedRunningTime="2026-02-17 00:09:53.478587576 +0000 UTC m=+267.214300367" Feb 17 00:09:53 crc kubenswrapper[4690]: I0217 00:09:53.478885 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qnd7w" podStartSLOduration=32.05423012 podStartE2EDuration="34.478876114s" podCreationTimestamp="2026-02-17 00:09:19 +0000 UTC" firstStartedPulling="2026-02-17 00:09:20.132502049 +0000 UTC m=+233.868214800" lastFinishedPulling="2026-02-17 00:09:22.557148043 +0000 UTC m=+236.292860794" observedRunningTime="2026-02-17 00:09:43.020616314 +0000 UTC m=+256.756329065" watchObservedRunningTime="2026-02-17 00:09:53.478876114 +0000 UTC m=+267.214588905" Feb 17 00:09:53 crc kubenswrapper[4690]: I0217 00:09:53.482832 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4glwj" podStartSLOduration=29.998829243 podStartE2EDuration="33.482818196s" podCreationTimestamp="2026-02-17 00:09:20 +0000 UTC" firstStartedPulling="2026-02-17 00:09:22.165447181 +0000 UTC m=+235.901159932" lastFinishedPulling="2026-02-17 00:09:25.649436124 +0000 UTC m=+239.385148885" observedRunningTime="2026-02-17 00:09:42.985749432 +0000 UTC m=+256.721462183" watchObservedRunningTime="2026-02-17 00:09:53.482818196 +0000 UTC m=+267.218530987" Feb 17 00:09:53 crc kubenswrapper[4690]: I0217 00:09:53.483764 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 17 00:09:53 crc kubenswrapper[4690]: I0217 00:09:53.483865 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 17 00:09:53 crc kubenswrapper[4690]: I0217 00:09:53.490242 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 00:09:53 crc kubenswrapper[4690]: I0217 00:09:53.528745 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=10.528716397 podStartE2EDuration="10.528716397s" podCreationTimestamp="2026-02-17 00:09:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:09:53.505223976 +0000 UTC m=+267.240936797" watchObservedRunningTime="2026-02-17 00:09:53.528716397 +0000 UTC m=+267.264429158" Feb 17 00:09:53 crc kubenswrapper[4690]: I0217 00:09:53.536998 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 17 00:09:54 crc kubenswrapper[4690]: I0217 00:09:54.372210 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 17 00:09:54 crc kubenswrapper[4690]: I0217 00:09:54.394411 4690 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 17 00:09:54 crc kubenswrapper[4690]: I0217 00:09:54.394736 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://32fe2039e895ae61d9818f5ae9926b9200de3bf091b06e767fc5568b9e3148d0" gracePeriod=5 Feb 17 00:09:54 crc kubenswrapper[4690]: I0217 00:09:54.407486 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 17 00:09:54 crc kubenswrapper[4690]: I0217 00:09:54.463948 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 17 00:09:54 crc kubenswrapper[4690]: I0217 00:09:54.811543 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 17 00:09:54 crc kubenswrapper[4690]: I0217 00:09:54.860395 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 17 00:09:54 crc kubenswrapper[4690]: I0217 00:09:54.900207 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.267014 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.293456 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.430521 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.468321 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.495995 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.524678 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.558642 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.641291 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.760738 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.761270 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.800033 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 17 00:09:55 crc kubenswrapper[4690]: I0217 00:09:55.886602 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.011076 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.137644 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.163001 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.361303 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.548603 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.656634 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.744231 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.764543 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.779510 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.839158 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.841244 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 17 00:09:56 crc kubenswrapper[4690]: I0217 00:09:56.870381 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.016582 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.030437 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.032555 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.067851 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.078619 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.266114 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.353034 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.381515 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.483691 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.488257 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.553423 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.591612 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.595912 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.618412 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.634419 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.643275 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.652806 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.748559 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.814347 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.848288 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.892155 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.941270 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 17 00:09:57 crc kubenswrapper[4690]: I0217 00:09:57.978923 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.102355 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.236417 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.272571 4690 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.298852 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.308430 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.377521 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.586261 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.588736 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.601479 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.630731 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.689464 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.705892 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.809143 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.854021 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 17 00:09:58 crc kubenswrapper[4690]: I0217 00:09:58.896338 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.030497 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.070869 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.080763 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.093116 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.117607 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.152611 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.196869 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.241811 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.264665 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.319092 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.343912 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.344143 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.432215 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.437751 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.448508 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.448575 4690 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="32fe2039e895ae61d9818f5ae9926b9200de3bf091b06e767fc5568b9e3148d0" exitCode=137 Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.519280 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.540734 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.592107 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.731910 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.859458 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.863047 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.883658 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.985207 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.985287 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:09:59 crc kubenswrapper[4690]: I0217 00:09:59.994374 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.062427 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.121599 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.129502 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.129832 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.130085 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.130322 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.130582 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.129651 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.129927 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.130457 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.130738 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.131716 4690 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.131953 4690 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.132120 4690 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.132261 4690 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.139132 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.145717 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.184850 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.217143 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.233513 4690 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.250473 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.312041 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.328776 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.338671 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.338763 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.360869 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.380946 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.426884 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.444154 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.457097 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.457157 4690 scope.go:117] "RemoveContainer" containerID="32fe2039e895ae61d9818f5ae9926b9200de3bf091b06e767fc5568b9e3148d0" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.457275 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.509284 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.528076 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.541975 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.575985 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.614772 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.685471 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.686130 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.689236 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.804476 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.805031 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.907270 4690 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.939545 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.953515 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 17 00:10:00 crc kubenswrapper[4690]: I0217 00:10:00.988197 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.058950 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.083249 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.086276 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.111674 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.122206 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.149086 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.228065 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.267054 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.273323 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.297612 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.303533 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.312059 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.325703 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.334156 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.379945 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.407248 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.407258 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.416706 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.469989 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.490935 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.606942 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.680700 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.720189 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.721217 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.728142 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.771399 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.809082 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.821284 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.922207 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 17 00:10:01 crc kubenswrapper[4690]: I0217 00:10:01.941517 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.081019 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.089068 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.094287 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.153012 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.281315 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.283945 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.286735 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.308036 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.320424 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.420243 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.525262 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.631325 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.648131 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.721548 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.777421 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.820393 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.897729 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 17 00:10:02 crc kubenswrapper[4690]: I0217 00:10:02.966064 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.064583 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.071793 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.072864 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.090898 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.185414 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.472658 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.472687 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.492066 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.497644 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.754046 4690 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.780046 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.825546 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.892612 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 17 00:10:03 crc kubenswrapper[4690]: I0217 00:10:03.937921 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.000822 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.020179 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.116500 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.228263 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.484114 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.511701 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.684664 4690 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.686434 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.769757 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.800651 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 17 00:10:04 crc kubenswrapper[4690]: I0217 00:10:04.961839 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.089456 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.130532 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.173113 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.225095 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.247605 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.283065 4690 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.389265 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.408807 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.478392 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.542700 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.640309 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.707440 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.731056 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.772126 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.796424 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.796971 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.879356 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.881626 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.922770 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.988336 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 17 00:10:05 crc kubenswrapper[4690]: I0217 00:10:05.989270 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.062394 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.112192 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.230517 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.235553 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.341318 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.491677 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.641323 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.671010 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.734075 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.740583 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.942182 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 17 00:10:06 crc kubenswrapper[4690]: I0217 00:10:06.971032 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.026236 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.221830 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.285471 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.310649 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.340134 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.372868 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.427334 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.486809 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.512874 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.517146 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.522422 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.763837 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 17 00:10:07 crc kubenswrapper[4690]: I0217 00:10:07.896345 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 17 00:10:08 crc kubenswrapper[4690]: I0217 00:10:08.198745 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 17 00:10:08 crc kubenswrapper[4690]: I0217 00:10:08.288774 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 17 00:10:08 crc kubenswrapper[4690]: I0217 00:10:08.447560 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 17 00:10:08 crc kubenswrapper[4690]: I0217 00:10:08.576244 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 17 00:10:08 crc kubenswrapper[4690]: I0217 00:10:08.729429 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 17 00:10:08 crc kubenswrapper[4690]: I0217 00:10:08.833732 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 17 00:10:09 crc kubenswrapper[4690]: I0217 00:10:09.227414 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 17 00:10:09 crc kubenswrapper[4690]: I0217 00:10:09.417565 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 17 00:10:09 crc kubenswrapper[4690]: I0217 00:10:09.560103 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 17 00:10:09 crc kubenswrapper[4690]: I0217 00:10:09.800701 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 17 00:10:10 crc kubenswrapper[4690]: I0217 00:10:10.113772 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 17 00:10:10 crc kubenswrapper[4690]: I0217 00:10:10.243799 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 17 00:10:10 crc kubenswrapper[4690]: I0217 00:10:10.826998 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 17 00:10:10 crc kubenswrapper[4690]: I0217 00:10:10.864980 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 17 00:10:26 crc kubenswrapper[4690]: I0217 00:10:26.762990 4690 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 17 00:10:33 crc kubenswrapper[4690]: I0217 00:10:33.739980 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657"] Feb 17 00:10:33 crc kubenswrapper[4690]: I0217 00:10:33.740733 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" podUID="159fd3d3-b560-4f05-b24e-25611073d6b8" containerName="route-controller-manager" containerID="cri-o://4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a" gracePeriod=30 Feb 17 00:10:33 crc kubenswrapper[4690]: I0217 00:10:33.745585 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gzx5v"] Feb 17 00:10:33 crc kubenswrapper[4690]: I0217 00:10:33.745778 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" podUID="8daa0570-a319-44d6-be28-c5717cb07575" containerName="controller-manager" containerID="cri-o://47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d" gracePeriod=30 Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.157318 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.164594 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.256662 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-client-ca\") pod \"8daa0570-a319-44d6-be28-c5717cb07575\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.256731 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xt7lg\" (UniqueName: \"kubernetes.io/projected/8daa0570-a319-44d6-be28-c5717cb07575-kube-api-access-xt7lg\") pod \"8daa0570-a319-44d6-be28-c5717cb07575\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.256777 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-proxy-ca-bundles\") pod \"8daa0570-a319-44d6-be28-c5717cb07575\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.256860 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8daa0570-a319-44d6-be28-c5717cb07575-serving-cert\") pod \"8daa0570-a319-44d6-be28-c5717cb07575\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.256880 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-config\") pod \"8daa0570-a319-44d6-be28-c5717cb07575\" (UID: \"8daa0570-a319-44d6-be28-c5717cb07575\") " Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.257100 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-client-ca" (OuterVolumeSpecName: "client-ca") pod "8daa0570-a319-44d6-be28-c5717cb07575" (UID: "8daa0570-a319-44d6-be28-c5717cb07575"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.257321 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.257375 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8daa0570-a319-44d6-be28-c5717cb07575" (UID: "8daa0570-a319-44d6-be28-c5717cb07575"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.258866 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-config" (OuterVolumeSpecName: "config") pod "8daa0570-a319-44d6-be28-c5717cb07575" (UID: "8daa0570-a319-44d6-be28-c5717cb07575"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.264019 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8daa0570-a319-44d6-be28-c5717cb07575-kube-api-access-xt7lg" (OuterVolumeSpecName: "kube-api-access-xt7lg") pod "8daa0570-a319-44d6-be28-c5717cb07575" (UID: "8daa0570-a319-44d6-be28-c5717cb07575"). InnerVolumeSpecName "kube-api-access-xt7lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.264220 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8daa0570-a319-44d6-be28-c5717cb07575-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8daa0570-a319-44d6-be28-c5717cb07575" (UID: "8daa0570-a319-44d6-be28-c5717cb07575"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.358132 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h7f6\" (UniqueName: \"kubernetes.io/projected/159fd3d3-b560-4f05-b24e-25611073d6b8-kube-api-access-6h7f6\") pod \"159fd3d3-b560-4f05-b24e-25611073d6b8\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.358221 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-client-ca\") pod \"159fd3d3-b560-4f05-b24e-25611073d6b8\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.358300 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-config\") pod \"159fd3d3-b560-4f05-b24e-25611073d6b8\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.358340 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159fd3d3-b560-4f05-b24e-25611073d6b8-serving-cert\") pod \"159fd3d3-b560-4f05-b24e-25611073d6b8\" (UID: \"159fd3d3-b560-4f05-b24e-25611073d6b8\") " Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.358539 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8daa0570-a319-44d6-be28-c5717cb07575-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.358552 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.358561 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xt7lg\" (UniqueName: \"kubernetes.io/projected/8daa0570-a319-44d6-be28-c5717cb07575-kube-api-access-xt7lg\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.358570 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8daa0570-a319-44d6-be28-c5717cb07575-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.359251 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-client-ca" (OuterVolumeSpecName: "client-ca") pod "159fd3d3-b560-4f05-b24e-25611073d6b8" (UID: "159fd3d3-b560-4f05-b24e-25611073d6b8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.359409 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-config" (OuterVolumeSpecName: "config") pod "159fd3d3-b560-4f05-b24e-25611073d6b8" (UID: "159fd3d3-b560-4f05-b24e-25611073d6b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.361416 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/159fd3d3-b560-4f05-b24e-25611073d6b8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "159fd3d3-b560-4f05-b24e-25611073d6b8" (UID: "159fd3d3-b560-4f05-b24e-25611073d6b8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.363556 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/159fd3d3-b560-4f05-b24e-25611073d6b8-kube-api-access-6h7f6" (OuterVolumeSpecName: "kube-api-access-6h7f6") pod "159fd3d3-b560-4f05-b24e-25611073d6b8" (UID: "159fd3d3-b560-4f05-b24e-25611073d6b8"). InnerVolumeSpecName "kube-api-access-6h7f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.460305 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.460347 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159fd3d3-b560-4f05-b24e-25611073d6b8-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.460372 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159fd3d3-b560-4f05-b24e-25611073d6b8-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.460385 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h7f6\" (UniqueName: \"kubernetes.io/projected/159fd3d3-b560-4f05-b24e-25611073d6b8-kube-api-access-6h7f6\") on node \"crc\" DevicePath \"\"" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.672803 4690 generic.go:334] "Generic (PLEG): container finished" podID="8daa0570-a319-44d6-be28-c5717cb07575" containerID="47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d" exitCode=0 Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.672922 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" event={"ID":"8daa0570-a319-44d6-be28-c5717cb07575","Type":"ContainerDied","Data":"47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d"} Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.673455 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" event={"ID":"8daa0570-a319-44d6-be28-c5717cb07575","Type":"ContainerDied","Data":"2a395ebd5fa6c242d817997f09e61db1fa0daf7dc0880751e34f5a32373d8350"} Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.673508 4690 scope.go:117] "RemoveContainer" containerID="47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.672988 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gzx5v" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.679112 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.679205 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" event={"ID":"159fd3d3-b560-4f05-b24e-25611073d6b8","Type":"ContainerDied","Data":"4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a"} Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.678996 4690 generic.go:334] "Generic (PLEG): container finished" podID="159fd3d3-b560-4f05-b24e-25611073d6b8" containerID="4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a" exitCode=0 Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.685843 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657" event={"ID":"159fd3d3-b560-4f05-b24e-25611073d6b8","Type":"ContainerDied","Data":"baac3fc17b094768715a02a9e059608c769d8977b936d108b1095394d167cae3"} Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.702532 4690 scope.go:117] "RemoveContainer" containerID="47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d" Feb 17 00:10:34 crc kubenswrapper[4690]: E0217 00:10:34.703095 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d\": container with ID starting with 47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d not found: ID does not exist" containerID="47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.703170 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d"} err="failed to get container status \"47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d\": rpc error: code = NotFound desc = could not find container \"47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d\": container with ID starting with 47079c8253071a3cf44a7bcf269f0b6dcc0e1331d0b92c879f6d9057c6e2989d not found: ID does not exist" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.703219 4690 scope.go:117] "RemoveContainer" containerID="4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.736832 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gzx5v"] Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.745065 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gzx5v"] Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.748904 4690 scope.go:117] "RemoveContainer" containerID="4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a" Feb 17 00:10:34 crc kubenswrapper[4690]: E0217 00:10:34.749475 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a\": container with ID starting with 4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a not found: ID does not exist" containerID="4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.749546 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a"} err="failed to get container status \"4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a\": rpc error: code = NotFound desc = could not find container \"4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a\": container with ID starting with 4fb7c9604f88b48e4afa5d57c90b4755c6396d46e6387c799f2ecc0bf604ec3a not found: ID does not exist" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.751021 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657"] Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.755565 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zw657"] Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.985107 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="159fd3d3-b560-4f05-b24e-25611073d6b8" path="/var/lib/kubelet/pods/159fd3d3-b560-4f05-b24e-25611073d6b8/volumes" Feb 17 00:10:34 crc kubenswrapper[4690]: I0217 00:10:34.986209 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8daa0570-a319-44d6-be28-c5717cb07575" path="/var/lib/kubelet/pods/8daa0570-a319-44d6-be28-c5717cb07575/volumes" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.248427 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7456d9bc85-96nw2"] Feb 17 00:10:35 crc kubenswrapper[4690]: E0217 00:10:35.248728 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8daa0570-a319-44d6-be28-c5717cb07575" containerName="controller-manager" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.248749 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8daa0570-a319-44d6-be28-c5717cb07575" containerName="controller-manager" Feb 17 00:10:35 crc kubenswrapper[4690]: E0217 00:10:35.248782 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" containerName="installer" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.248796 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" containerName="installer" Feb 17 00:10:35 crc kubenswrapper[4690]: E0217 00:10:35.248817 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="159fd3d3-b560-4f05-b24e-25611073d6b8" containerName="route-controller-manager" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.248831 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="159fd3d3-b560-4f05-b24e-25611073d6b8" containerName="route-controller-manager" Feb 17 00:10:35 crc kubenswrapper[4690]: E0217 00:10:35.248854 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.248867 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.249020 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.249064 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="8daa0570-a319-44d6-be28-c5717cb07575" containerName="controller-manager" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.249081 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d36eab40-817c-4939-a0d6-ea1fbb2949ca" containerName="installer" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.249101 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="159fd3d3-b560-4f05-b24e-25611073d6b8" containerName="route-controller-manager" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.249673 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.268505 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.270431 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.270829 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.271560 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.272072 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.274847 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.284111 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.295396 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z"] Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.296259 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.300008 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.300561 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.300766 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.301352 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.304851 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.305799 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.315648 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z"] Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.326287 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7456d9bc85-96nw2"] Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.378140 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-config\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.378220 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/faaf5781-4486-45ea-b50c-9e9a24069dc3-serving-cert\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.378413 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-client-ca\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.378510 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhjkv\" (UniqueName: \"kubernetes.io/projected/faaf5781-4486-45ea-b50c-9e9a24069dc3-kube-api-access-zhjkv\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.378580 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-proxy-ca-bundles\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.479723 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05f378da-f53f-4164-976f-d719703a2a95-serving-cert\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.479983 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05f378da-f53f-4164-976f-d719703a2a95-config\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.480073 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzssm\" (UniqueName: \"kubernetes.io/projected/05f378da-f53f-4164-976f-d719703a2a95-kube-api-access-wzssm\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.480118 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-config\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.480155 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05f378da-f53f-4164-976f-d719703a2a95-client-ca\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.480197 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/faaf5781-4486-45ea-b50c-9e9a24069dc3-serving-cert\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.480320 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-client-ca\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.480456 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhjkv\" (UniqueName: \"kubernetes.io/projected/faaf5781-4486-45ea-b50c-9e9a24069dc3-kube-api-access-zhjkv\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.480538 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-proxy-ca-bundles\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.481916 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-config\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.482269 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-client-ca\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.482294 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-proxy-ca-bundles\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.499956 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/faaf5781-4486-45ea-b50c-9e9a24069dc3-serving-cert\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.500781 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhjkv\" (UniqueName: \"kubernetes.io/projected/faaf5781-4486-45ea-b50c-9e9a24069dc3-kube-api-access-zhjkv\") pod \"controller-manager-7456d9bc85-96nw2\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.582403 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05f378da-f53f-4164-976f-d719703a2a95-serving-cert\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.582809 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05f378da-f53f-4164-976f-d719703a2a95-config\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.583060 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzssm\" (UniqueName: \"kubernetes.io/projected/05f378da-f53f-4164-976f-d719703a2a95-kube-api-access-wzssm\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.583283 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05f378da-f53f-4164-976f-d719703a2a95-client-ca\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.584848 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05f378da-f53f-4164-976f-d719703a2a95-config\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.585355 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05f378da-f53f-4164-976f-d719703a2a95-client-ca\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.596082 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.597154 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05f378da-f53f-4164-976f-d719703a2a95-serving-cert\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.601987 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzssm\" (UniqueName: \"kubernetes.io/projected/05f378da-f53f-4164-976f-d719703a2a95-kube-api-access-wzssm\") pod \"route-controller-manager-64975bdc4c-dfq7z\" (UID: \"05f378da-f53f-4164-976f-d719703a2a95\") " pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:35 crc kubenswrapper[4690]: I0217 00:10:35.621821 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.049709 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z"] Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.071723 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7456d9bc85-96nw2"] Feb 17 00:10:36 crc kubenswrapper[4690]: W0217 00:10:36.075763 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfaaf5781_4486_45ea_b50c_9e9a24069dc3.slice/crio-3d36d7f4271e102c3a91fa1db875a97c48a63af9b0392ebdbcbca7d45c7a31d0 WatchSource:0}: Error finding container 3d36d7f4271e102c3a91fa1db875a97c48a63af9b0392ebdbcbca7d45c7a31d0: Status 404 returned error can't find the container with id 3d36d7f4271e102c3a91fa1db875a97c48a63af9b0392ebdbcbca7d45c7a31d0 Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.701556 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" event={"ID":"05f378da-f53f-4164-976f-d719703a2a95","Type":"ContainerStarted","Data":"df8859ff2cc0b9b138bd1d8e574b48022f44f6930d59403ac909cd4ad750ad79"} Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.701979 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.702010 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" event={"ID":"05f378da-f53f-4164-976f-d719703a2a95","Type":"ContainerStarted","Data":"050a7ca1f31df957b232bfd1f2992bb8cdb919e71505efdeb2cecb6c57a1482c"} Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.703080 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" event={"ID":"faaf5781-4486-45ea-b50c-9e9a24069dc3","Type":"ContainerStarted","Data":"efd434c9f9d7b4f0208dab9a4f7c8f9bdf6f279850b9cb6ffdd7c00e86912697"} Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.703106 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" event={"ID":"faaf5781-4486-45ea-b50c-9e9a24069dc3","Type":"ContainerStarted","Data":"3d36d7f4271e102c3a91fa1db875a97c48a63af9b0392ebdbcbca7d45c7a31d0"} Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.703374 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.708317 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.712323 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" Feb 17 00:10:36 crc kubenswrapper[4690]: I0217 00:10:36.721491 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-64975bdc4c-dfq7z" podStartSLOduration=3.721472082 podStartE2EDuration="3.721472082s" podCreationTimestamp="2026-02-17 00:10:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:10:36.718392684 +0000 UTC m=+310.454105435" watchObservedRunningTime="2026-02-17 00:10:36.721472082 +0000 UTC m=+310.457184833" Feb 17 00:10:56 crc kubenswrapper[4690]: I0217 00:10:56.325549 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:10:56 crc kubenswrapper[4690]: I0217 00:10:56.326158 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:11:26 crc kubenswrapper[4690]: I0217 00:11:26.326154 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:11:26 crc kubenswrapper[4690]: I0217 00:11:26.326836 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:11:33 crc kubenswrapper[4690]: I0217 00:11:33.743930 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" podStartSLOduration=60.74391161 podStartE2EDuration="1m0.74391161s" podCreationTimestamp="2026-02-17 00:10:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:10:36.774718793 +0000 UTC m=+310.510431544" watchObservedRunningTime="2026-02-17 00:11:33.74391161 +0000 UTC m=+367.479624371" Feb 17 00:11:33 crc kubenswrapper[4690]: I0217 00:11:33.745748 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7456d9bc85-96nw2"] Feb 17 00:11:33 crc kubenswrapper[4690]: I0217 00:11:33.745968 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" podUID="faaf5781-4486-45ea-b50c-9e9a24069dc3" containerName="controller-manager" containerID="cri-o://efd434c9f9d7b4f0208dab9a4f7c8f9bdf6f279850b9cb6ffdd7c00e86912697" gracePeriod=30 Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.038547 4690 generic.go:334] "Generic (PLEG): container finished" podID="faaf5781-4486-45ea-b50c-9e9a24069dc3" containerID="efd434c9f9d7b4f0208dab9a4f7c8f9bdf6f279850b9cb6ffdd7c00e86912697" exitCode=0 Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.038586 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" event={"ID":"faaf5781-4486-45ea-b50c-9e9a24069dc3","Type":"ContainerDied","Data":"efd434c9f9d7b4f0208dab9a4f7c8f9bdf6f279850b9cb6ffdd7c00e86912697"} Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.134069 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.165226 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhjkv\" (UniqueName: \"kubernetes.io/projected/faaf5781-4486-45ea-b50c-9e9a24069dc3-kube-api-access-zhjkv\") pod \"faaf5781-4486-45ea-b50c-9e9a24069dc3\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.165628 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-client-ca\") pod \"faaf5781-4486-45ea-b50c-9e9a24069dc3\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.165708 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-proxy-ca-bundles\") pod \"faaf5781-4486-45ea-b50c-9e9a24069dc3\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.165738 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/faaf5781-4486-45ea-b50c-9e9a24069dc3-serving-cert\") pod \"faaf5781-4486-45ea-b50c-9e9a24069dc3\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.165785 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-config\") pod \"faaf5781-4486-45ea-b50c-9e9a24069dc3\" (UID: \"faaf5781-4486-45ea-b50c-9e9a24069dc3\") " Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.166834 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "faaf5781-4486-45ea-b50c-9e9a24069dc3" (UID: "faaf5781-4486-45ea-b50c-9e9a24069dc3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.166914 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-config" (OuterVolumeSpecName: "config") pod "faaf5781-4486-45ea-b50c-9e9a24069dc3" (UID: "faaf5781-4486-45ea-b50c-9e9a24069dc3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.166955 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-client-ca" (OuterVolumeSpecName: "client-ca") pod "faaf5781-4486-45ea-b50c-9e9a24069dc3" (UID: "faaf5781-4486-45ea-b50c-9e9a24069dc3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.171693 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faaf5781-4486-45ea-b50c-9e9a24069dc3-kube-api-access-zhjkv" (OuterVolumeSpecName: "kube-api-access-zhjkv") pod "faaf5781-4486-45ea-b50c-9e9a24069dc3" (UID: "faaf5781-4486-45ea-b50c-9e9a24069dc3"). InnerVolumeSpecName "kube-api-access-zhjkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.171806 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faaf5781-4486-45ea-b50c-9e9a24069dc3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "faaf5781-4486-45ea-b50c-9e9a24069dc3" (UID: "faaf5781-4486-45ea-b50c-9e9a24069dc3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.267764 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.267818 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/faaf5781-4486-45ea-b50c-9e9a24069dc3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.267837 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.267857 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhjkv\" (UniqueName: \"kubernetes.io/projected/faaf5781-4486-45ea-b50c-9e9a24069dc3-kube-api-access-zhjkv\") on node \"crc\" DevicePath \"\"" Feb 17 00:11:34 crc kubenswrapper[4690]: I0217 00:11:34.267879 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/faaf5781-4486-45ea-b50c-9e9a24069dc3-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.047680 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" event={"ID":"faaf5781-4486-45ea-b50c-9e9a24069dc3","Type":"ContainerDied","Data":"3d36d7f4271e102c3a91fa1db875a97c48a63af9b0392ebdbcbca7d45c7a31d0"} Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.047742 4690 scope.go:117] "RemoveContainer" containerID="efd434c9f9d7b4f0208dab9a4f7c8f9bdf6f279850b9cb6ffdd7c00e86912697" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.047747 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7456d9bc85-96nw2" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.073227 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7456d9bc85-96nw2"] Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.078066 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7456d9bc85-96nw2"] Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.296624 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-766bdf997d-6pzb9"] Feb 17 00:11:35 crc kubenswrapper[4690]: E0217 00:11:35.296846 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faaf5781-4486-45ea-b50c-9e9a24069dc3" containerName="controller-manager" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.296861 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="faaf5781-4486-45ea-b50c-9e9a24069dc3" containerName="controller-manager" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.297004 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="faaf5781-4486-45ea-b50c-9e9a24069dc3" containerName="controller-manager" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.297502 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.302865 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.304339 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.304493 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.304561 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.304506 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.304900 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.312523 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-766bdf997d-6pzb9"] Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.313442 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.388892 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6730484b-777c-4862-81d2-41be42a0ab9b-serving-cert\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.389064 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmn4z\" (UniqueName: \"kubernetes.io/projected/6730484b-777c-4862-81d2-41be42a0ab9b-kube-api-access-tmn4z\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.389212 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6730484b-777c-4862-81d2-41be42a0ab9b-proxy-ca-bundles\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.389263 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6730484b-777c-4862-81d2-41be42a0ab9b-config\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.389562 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6730484b-777c-4862-81d2-41be42a0ab9b-client-ca\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.490502 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6730484b-777c-4862-81d2-41be42a0ab9b-proxy-ca-bundles\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.490578 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6730484b-777c-4862-81d2-41be42a0ab9b-config\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.490633 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6730484b-777c-4862-81d2-41be42a0ab9b-client-ca\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.490695 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6730484b-777c-4862-81d2-41be42a0ab9b-serving-cert\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.490739 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmn4z\" (UniqueName: \"kubernetes.io/projected/6730484b-777c-4862-81d2-41be42a0ab9b-kube-api-access-tmn4z\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.492106 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6730484b-777c-4862-81d2-41be42a0ab9b-client-ca\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.492297 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6730484b-777c-4862-81d2-41be42a0ab9b-config\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.492627 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6730484b-777c-4862-81d2-41be42a0ab9b-proxy-ca-bundles\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.500607 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6730484b-777c-4862-81d2-41be42a0ab9b-serving-cert\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.507496 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmn4z\" (UniqueName: \"kubernetes.io/projected/6730484b-777c-4862-81d2-41be42a0ab9b-kube-api-access-tmn4z\") pod \"controller-manager-766bdf997d-6pzb9\" (UID: \"6730484b-777c-4862-81d2-41be42a0ab9b\") " pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.626231 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:35 crc kubenswrapper[4690]: I0217 00:11:35.914536 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-766bdf997d-6pzb9"] Feb 17 00:11:36 crc kubenswrapper[4690]: I0217 00:11:36.057978 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" event={"ID":"6730484b-777c-4862-81d2-41be42a0ab9b","Type":"ContainerStarted","Data":"352da69f657fb589c6a9047ce880e9c137c4d8ef511ff8fd7c1f76b2f6bdb73d"} Feb 17 00:11:36 crc kubenswrapper[4690]: I0217 00:11:36.058050 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" event={"ID":"6730484b-777c-4862-81d2-41be42a0ab9b","Type":"ContainerStarted","Data":"e9bc1f75c4e7bcc4fcd9a6ebfc2d20b69a22b7db3b59ef379af13578d823dac4"} Feb 17 00:11:36 crc kubenswrapper[4690]: I0217 00:11:36.058324 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:36 crc kubenswrapper[4690]: I0217 00:11:36.060836 4690 patch_prober.go:28] interesting pod/controller-manager-766bdf997d-6pzb9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Feb 17 00:11:36 crc kubenswrapper[4690]: I0217 00:11:36.060972 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" podUID="6730484b-777c-4862-81d2-41be42a0ab9b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Feb 17 00:11:36 crc kubenswrapper[4690]: I0217 00:11:36.083048 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" podStartSLOduration=3.083023379 podStartE2EDuration="3.083023379s" podCreationTimestamp="2026-02-17 00:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:11:36.081501016 +0000 UTC m=+369.817213777" watchObservedRunningTime="2026-02-17 00:11:36.083023379 +0000 UTC m=+369.818736160" Feb 17 00:11:36 crc kubenswrapper[4690]: I0217 00:11:36.983669 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faaf5781-4486-45ea-b50c-9e9a24069dc3" path="/var/lib/kubelet/pods/faaf5781-4486-45ea-b50c-9e9a24069dc3/volumes" Feb 17 00:11:37 crc kubenswrapper[4690]: I0217 00:11:37.074702 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-766bdf997d-6pzb9" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.100077 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-86jzf"] Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.102271 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.146718 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/596d3087-c020-4308-8123-fc37e971bfa6-registry-certificates\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.146797 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-765tc\" (UniqueName: \"kubernetes.io/projected/596d3087-c020-4308-8123-fc37e971bfa6-kube-api-access-765tc\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.146833 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.146872 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/596d3087-c020-4308-8123-fc37e971bfa6-bound-sa-token\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.146902 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/596d3087-c020-4308-8123-fc37e971bfa6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.146971 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/596d3087-c020-4308-8123-fc37e971bfa6-trusted-ca\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.146995 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/596d3087-c020-4308-8123-fc37e971bfa6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.147030 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/596d3087-c020-4308-8123-fc37e971bfa6-registry-tls\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.148090 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-86jzf"] Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.190538 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.248567 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/596d3087-c020-4308-8123-fc37e971bfa6-registry-certificates\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.248638 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-765tc\" (UniqueName: \"kubernetes.io/projected/596d3087-c020-4308-8123-fc37e971bfa6-kube-api-access-765tc\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.248682 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/596d3087-c020-4308-8123-fc37e971bfa6-bound-sa-token\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.248715 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/596d3087-c020-4308-8123-fc37e971bfa6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.248746 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/596d3087-c020-4308-8123-fc37e971bfa6-trusted-ca\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.248766 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/596d3087-c020-4308-8123-fc37e971bfa6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.248799 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/596d3087-c020-4308-8123-fc37e971bfa6-registry-tls\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.250432 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/596d3087-c020-4308-8123-fc37e971bfa6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.251148 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/596d3087-c020-4308-8123-fc37e971bfa6-trusted-ca\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.252057 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/596d3087-c020-4308-8123-fc37e971bfa6-registry-certificates\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.256570 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/596d3087-c020-4308-8123-fc37e971bfa6-registry-tls\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.256892 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/596d3087-c020-4308-8123-fc37e971bfa6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.290038 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/596d3087-c020-4308-8123-fc37e971bfa6-bound-sa-token\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.291943 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-765tc\" (UniqueName: \"kubernetes.io/projected/596d3087-c020-4308-8123-fc37e971bfa6-kube-api-access-765tc\") pod \"image-registry-66df7c8f76-86jzf\" (UID: \"596d3087-c020-4308-8123-fc37e971bfa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.433903 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:46 crc kubenswrapper[4690]: I0217 00:11:46.934775 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-86jzf"] Feb 17 00:11:47 crc kubenswrapper[4690]: I0217 00:11:47.130076 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" event={"ID":"596d3087-c020-4308-8123-fc37e971bfa6","Type":"ContainerStarted","Data":"0b052096078e7014f6ef31dde201eedc0017564777239cd2d50b60971008a856"} Feb 17 00:11:47 crc kubenswrapper[4690]: I0217 00:11:47.130474 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" event={"ID":"596d3087-c020-4308-8123-fc37e971bfa6","Type":"ContainerStarted","Data":"2d46f796bb4ebbe881892456b1fe6da05ea104fd467d20bea0022153d174863b"} Feb 17 00:11:47 crc kubenswrapper[4690]: I0217 00:11:47.130498 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:11:47 crc kubenswrapper[4690]: I0217 00:11:47.155870 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" podStartSLOduration=1.155839541 podStartE2EDuration="1.155839541s" podCreationTimestamp="2026-02-17 00:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:11:47.151655962 +0000 UTC m=+380.887368733" watchObservedRunningTime="2026-02-17 00:11:47.155839541 +0000 UTC m=+380.891552302" Feb 17 00:11:56 crc kubenswrapper[4690]: I0217 00:11:56.325711 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:11:56 crc kubenswrapper[4690]: I0217 00:11:56.326389 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:11:56 crc kubenswrapper[4690]: I0217 00:11:56.326457 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:11:56 crc kubenswrapper[4690]: I0217 00:11:56.327225 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"994b411b25ff62fc16205928af96a4f9d2be45f24e3023cf34d1411551126d69"} pod="openshift-machine-config-operator/machine-config-daemon-lszwm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 00:11:56 crc kubenswrapper[4690]: I0217 00:11:56.327309 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" containerID="cri-o://994b411b25ff62fc16205928af96a4f9d2be45f24e3023cf34d1411551126d69" gracePeriod=600 Feb 17 00:11:57 crc kubenswrapper[4690]: I0217 00:11:57.234681 4690 generic.go:334] "Generic (PLEG): container finished" podID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerID="994b411b25ff62fc16205928af96a4f9d2be45f24e3023cf34d1411551126d69" exitCode=0 Feb 17 00:11:57 crc kubenswrapper[4690]: I0217 00:11:57.234821 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerDied","Data":"994b411b25ff62fc16205928af96a4f9d2be45f24e3023cf34d1411551126d69"} Feb 17 00:11:57 crc kubenswrapper[4690]: I0217 00:11:57.235630 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerStarted","Data":"2bfdb84a832fb3eb0bc54fa1a1fa0b2de03c4db1987487e9015bfa4b3b7cb1bb"} Feb 17 00:11:57 crc kubenswrapper[4690]: I0217 00:11:57.235655 4690 scope.go:117] "RemoveContainer" containerID="e7da5461677ae6e63bd58ac210855c6bd2edf18c6fbe9b6ad7081296ee5a1de6" Feb 17 00:12:06 crc kubenswrapper[4690]: I0217 00:12:06.440560 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-86jzf" Feb 17 00:12:06 crc kubenswrapper[4690]: I0217 00:12:06.501861 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9pwzh"] Feb 17 00:12:27 crc kubenswrapper[4690]: I0217 00:12:27.140195 4690 scope.go:117] "RemoveContainer" containerID="f71a50accba2e48ecab8c5ff6a561e5a1174089552aa00cb1e826551157975ef" Feb 17 00:12:27 crc kubenswrapper[4690]: I0217 00:12:27.170675 4690 scope.go:117] "RemoveContainer" containerID="b5ba2f289d9a5f35a68aee0109b2015f984f6f86c0340b0dffecca84b51d224d" Feb 17 00:12:27 crc kubenswrapper[4690]: I0217 00:12:27.191576 4690 scope.go:117] "RemoveContainer" containerID="891ca0dc7d2966cd7617785901b62952bc2aa02df1b6f736ce145da417ed559b" Feb 17 00:12:27 crc kubenswrapper[4690]: I0217 00:12:27.208918 4690 scope.go:117] "RemoveContainer" containerID="3b1ee5617311d314d6cbe763d8ee904b86f450882adb56ddceaf8453ba9877e3" Feb 17 00:12:27 crc kubenswrapper[4690]: I0217 00:12:27.221806 4690 scope.go:117] "RemoveContainer" containerID="099f13f079f56ba9b6930622b3d36c56a24ec34486a1ff7c6c51ae0ea780f73d" Feb 17 00:12:27 crc kubenswrapper[4690]: I0217 00:12:27.236525 4690 scope.go:117] "RemoveContainer" containerID="2f822190e3a20aec4e82f6dc10a17fbd854d452b231a7c1fffa58389ef344a4b" Feb 17 00:12:31 crc kubenswrapper[4690]: I0217 00:12:31.546931 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" podUID="2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" containerName="registry" containerID="cri-o://41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b" gracePeriod=30 Feb 17 00:12:31 crc kubenswrapper[4690]: I0217 00:12:31.968210 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.084430 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-certificates\") pod \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.085552 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-bound-sa-token\") pod \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.085654 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.085620 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-trusted-ca\") pod \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.085785 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhdtv\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-kube-api-access-jhdtv\") pod \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.085827 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-tls\") pod \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.085867 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-installation-pull-secrets\") pod \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.086101 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.086205 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-ca-trust-extracted\") pod \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\" (UID: \"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e\") " Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.086319 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.087769 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.087804 4690 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.092338 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.092641 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.093147 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-kube-api-access-jhdtv" (OuterVolumeSpecName: "kube-api-access-jhdtv") pod "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e"). InnerVolumeSpecName "kube-api-access-jhdtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.098959 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.101715 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.103799 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" (UID: "2cc659a3-51d1-4ebd-8cdc-42f6a648c65e"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.189118 4690 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.189164 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhdtv\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-kube-api-access-jhdtv\") on node \"crc\" DevicePath \"\"" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.189187 4690 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.189204 4690 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.189225 4690 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.480522 4690 generic.go:334] "Generic (PLEG): container finished" podID="2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" containerID="41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b" exitCode=0 Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.480778 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" event={"ID":"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e","Type":"ContainerDied","Data":"41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b"} Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.480833 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" event={"ID":"2cc659a3-51d1-4ebd-8cdc-42f6a648c65e","Type":"ContainerDied","Data":"55313a5fee0f3cd37f3a73b837c05d53ee3d188b579e8a870c2597370bd4939d"} Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.480862 4690 scope.go:117] "RemoveContainer" containerID="41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.481160 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9pwzh" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.519294 4690 scope.go:117] "RemoveContainer" containerID="41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b" Feb 17 00:12:32 crc kubenswrapper[4690]: E0217 00:12:32.520988 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b\": container with ID starting with 41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b not found: ID does not exist" containerID="41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.521098 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b"} err="failed to get container status \"41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b\": rpc error: code = NotFound desc = could not find container \"41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b\": container with ID starting with 41a6ecd9ae502e28af91c911a11adc24933b6c1911d6e8288ed3f3fa91db754b not found: ID does not exist" Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.522231 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9pwzh"] Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.526177 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9pwzh"] Feb 17 00:12:32 crc kubenswrapper[4690]: I0217 00:12:32.985695 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" path="/var/lib/kubelet/pods/2cc659a3-51d1-4ebd-8cdc-42f6a648c65e/volumes" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.556967 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2frwb"] Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.565765 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovn-controller" containerID="cri-o://a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869" gracePeriod=30 Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.565876 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="kube-rbac-proxy-node" containerID="cri-o://515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4" gracePeriod=30 Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.565891 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b" gracePeriod=30 Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.565802 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="nbdb" containerID="cri-o://a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616" gracePeriod=30 Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.565949 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovn-acl-logging" containerID="cri-o://deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf" gracePeriod=30 Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.565859 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="sbdb" containerID="cri-o://dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624" gracePeriod=30 Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.566025 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="northd" containerID="cri-o://5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8" gracePeriod=30 Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.603490 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" containerID="cri-o://e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65" gracePeriod=30 Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.865703 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/3.log" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.868538 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovn-acl-logging/0.log" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.869112 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovn-controller/0.log" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.869555 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.938863 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lmrvp"] Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939070 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="northd" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939083 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="northd" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939096 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939102 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939109 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="nbdb" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939116 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="nbdb" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939126 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovn-acl-logging" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939131 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovn-acl-logging" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939141 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="kube-rbac-proxy-node" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939146 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="kube-rbac-proxy-node" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939155 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="sbdb" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939161 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="sbdb" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939167 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="kubecfg-setup" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939172 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="kubecfg-setup" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939179 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939186 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939193 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939199 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939205 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" containerName="registry" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939212 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" containerName="registry" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939219 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939225 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939234 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovn-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939239 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovn-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939248 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="kube-rbac-proxy-ovn-metrics" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939254 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="kube-rbac-proxy-ovn-metrics" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939339 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cc659a3-51d1-4ebd-8cdc-42f6a648c65e" containerName="registry" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939352 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovn-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939363 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="northd" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939402 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939411 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="kube-rbac-proxy-ovn-metrics" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939418 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="nbdb" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939426 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939434 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovn-acl-logging" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939442 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939448 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="sbdb" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939455 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939464 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="kube-rbac-proxy-node" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939470 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: E0217 00:13:54.939572 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.939582 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerName="ovnkube-controller" Feb 17 00:13:54 crc kubenswrapper[4690]: I0217 00:13:54.941402 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.006847 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovn-node-metrics-cert\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.006899 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-env-overrides\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.006927 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-slash\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.006958 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqxq5\" (UniqueName: \"kubernetes.io/projected/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-kube-api-access-pqxq5\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.006984 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007010 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-openvswitch\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007040 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-etc-openvswitch\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007078 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-config\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007103 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-ovn-kubernetes\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007139 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-kubelet\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007168 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-netd\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007188 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-systemd\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007211 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-bin\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007234 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-var-lib-openvswitch\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007259 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-systemd-units\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007287 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-ovn\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007307 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-node-log\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007341 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-netns\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007391 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-log-socket\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007430 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-script-lib\") pod \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\" (UID: \"c6eeb20f-1ee2-4d96-a53b-da30d60213f3\") " Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007486 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007627 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007664 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-slash" (OuterVolumeSpecName: "host-slash") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007674 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007788 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-log-socket" (OuterVolumeSpecName: "log-socket") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007711 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007739 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007735 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007749 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-node-log" (OuterVolumeSpecName: "node-log") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007767 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007784 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007788 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.007694 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.008079 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.008413 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.008430 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.008966 4690 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.008995 4690 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009008 4690 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009022 4690 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009035 4690 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009048 4690 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009059 4690 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009071 4690 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009082 4690 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009093 4690 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009104 4690 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009114 4690 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-node-log\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009126 4690 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009137 4690 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-log-socket\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009149 4690 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.009161 4690 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-host-slash\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.010601 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.013119 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.016597 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-kube-api-access-pqxq5" (OuterVolumeSpecName: "kube-api-access-pqxq5") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "kube-api-access-pqxq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.025711 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "c6eeb20f-1ee2-4d96-a53b-da30d60213f3" (UID: "c6eeb20f-1ee2-4d96-a53b-da30d60213f3"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.049980 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovnkube-controller/3.log" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.052658 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovn-acl-logging/0.log" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053162 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2frwb_c6eeb20f-1ee2-4d96-a53b-da30d60213f3/ovn-controller/0.log" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053625 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65" exitCode=0 Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053669 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624" exitCode=0 Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053685 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616" exitCode=0 Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053696 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8" exitCode=0 Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053717 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b" exitCode=0 Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053728 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4" exitCode=0 Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053740 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf" exitCode=143 Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053751 4690 generic.go:334] "Generic (PLEG): container finished" podID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" containerID="a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869" exitCode=143 Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053757 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053667 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053820 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053833 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053844 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053875 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053886 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053900 4690 scope.go:117] "RemoveContainer" containerID="e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053897 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053922 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053962 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053971 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053977 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.053995 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054001 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054007 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054013 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054081 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054094 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054101 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054107 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054114 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054144 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054150 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054155 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054160 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054165 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054169 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054193 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054221 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054229 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054234 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054239 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054487 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054495 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054501 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054507 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054513 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054518 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054546 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2frwb" event={"ID":"c6eeb20f-1ee2-4d96-a53b-da30d60213f3","Type":"ContainerDied","Data":"8e53d779a45111fcc088635b7ab3cbba7b3db883020c91984e6348af43fab072"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054557 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054564 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054569 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054575 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054580 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054586 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054593 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054598 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054604 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.054623 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.055459 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/2.log" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.056154 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/1.log" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.056199 4690 generic.go:334] "Generic (PLEG): container finished" podID="797b14df-f7d0-419b-95f8-f02f25409e66" containerID="b9405af1a1a3ca87c5a1110233884fa8b8dad590f816aa1ac3af9d2e5dd32ff6" exitCode=2 Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.056231 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vb4v5" event={"ID":"797b14df-f7d0-419b-95f8-f02f25409e66","Type":"ContainerDied","Data":"b9405af1a1a3ca87c5a1110233884fa8b8dad590f816aa1ac3af9d2e5dd32ff6"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.056259 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae"} Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.056906 4690 scope.go:117] "RemoveContainer" containerID="b9405af1a1a3ca87c5a1110233884fa8b8dad590f816aa1ac3af9d2e5dd32ff6" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.057209 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vb4v5_openshift-multus(797b14df-f7d0-419b-95f8-f02f25409e66)\"" pod="openshift-multus/multus-vb4v5" podUID="797b14df-f7d0-419b-95f8-f02f25409e66" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.072796 4690 scope.go:117] "RemoveContainer" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.096582 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2frwb"] Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.102019 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2frwb"] Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110010 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-run-ovn\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110114 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-run-netns\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110159 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm9tn\" (UniqueName: \"kubernetes.io/projected/1332a26f-1afd-4af0-bde4-77186a1350b5-kube-api-access-cm9tn\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110191 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1332a26f-1afd-4af0-bde4-77186a1350b5-ovnkube-script-lib\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110227 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-node-log\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110280 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-cni-bin\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110313 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-var-lib-openvswitch\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110362 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110420 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-systemd-units\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110453 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1332a26f-1afd-4af0-bde4-77186a1350b5-env-overrides\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110491 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-cni-netd\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110519 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-run-ovn-kubernetes\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110546 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1332a26f-1afd-4af0-bde4-77186a1350b5-ovn-node-metrics-cert\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110575 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-etc-openvswitch\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110605 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1332a26f-1afd-4af0-bde4-77186a1350b5-ovnkube-config\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110637 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-kubelet\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110679 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-slash\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110705 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-log-socket\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110744 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-run-systemd\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110779 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-run-openvswitch\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110843 4690 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110863 4690 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110882 4690 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.110899 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqxq5\" (UniqueName: \"kubernetes.io/projected/c6eeb20f-1ee2-4d96-a53b-da30d60213f3-kube-api-access-pqxq5\") on node \"crc\" DevicePath \"\"" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.111419 4690 scope.go:117] "RemoveContainer" containerID="dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.126103 4690 scope.go:117] "RemoveContainer" containerID="a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.151836 4690 scope.go:117] "RemoveContainer" containerID="5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.167454 4690 scope.go:117] "RemoveContainer" containerID="486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.182317 4690 scope.go:117] "RemoveContainer" containerID="515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.193627 4690 scope.go:117] "RemoveContainer" containerID="deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.204325 4690 scope.go:117] "RemoveContainer" containerID="a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212105 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-run-ovn\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212150 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-run-netns\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212175 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm9tn\" (UniqueName: \"kubernetes.io/projected/1332a26f-1afd-4af0-bde4-77186a1350b5-kube-api-access-cm9tn\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212192 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1332a26f-1afd-4af0-bde4-77186a1350b5-ovnkube-script-lib\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212212 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-var-lib-openvswitch\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212226 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-node-log\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212244 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-cni-bin\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212264 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212282 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-systemd-units\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212297 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1332a26f-1afd-4af0-bde4-77186a1350b5-env-overrides\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212319 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-cni-netd\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212333 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-run-ovn-kubernetes\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212348 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1332a26f-1afd-4af0-bde4-77186a1350b5-ovn-node-metrics-cert\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212382 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1332a26f-1afd-4af0-bde4-77186a1350b5-ovnkube-config\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212399 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-etc-openvswitch\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212417 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-kubelet\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212449 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-slash\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212466 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-log-socket\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212486 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-run-systemd\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212513 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-run-openvswitch\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212604 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-run-ovn\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212652 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-slash\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212622 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-run-netns\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212716 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-run-systemd\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212742 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-log-socket\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.212768 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213126 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-node-log\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213170 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-var-lib-openvswitch\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213287 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-run-openvswitch\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213332 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-kubelet\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213361 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-etc-openvswitch\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213416 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1332a26f-1afd-4af0-bde4-77186a1350b5-ovnkube-config\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213468 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-cni-netd\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213502 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-run-ovn-kubernetes\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213516 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-systemd-units\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213535 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1332a26f-1afd-4af0-bde4-77186a1350b5-host-cni-bin\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213593 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1332a26f-1afd-4af0-bde4-77186a1350b5-ovnkube-script-lib\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.213595 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1332a26f-1afd-4af0-bde4-77186a1350b5-env-overrides\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.215029 4690 scope.go:117] "RemoveContainer" containerID="1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.217551 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1332a26f-1afd-4af0-bde4-77186a1350b5-ovn-node-metrics-cert\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.228211 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm9tn\" (UniqueName: \"kubernetes.io/projected/1332a26f-1afd-4af0-bde4-77186a1350b5-kube-api-access-cm9tn\") pod \"ovnkube-node-lmrvp\" (UID: \"1332a26f-1afd-4af0-bde4-77186a1350b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.228716 4690 scope.go:117] "RemoveContainer" containerID="e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.229010 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": container with ID starting with e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65 not found: ID does not exist" containerID="e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.229043 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65"} err="failed to get container status \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": rpc error: code = NotFound desc = could not find container \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": container with ID starting with e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.229068 4690 scope.go:117] "RemoveContainer" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.229342 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\": container with ID starting with e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340 not found: ID does not exist" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.229383 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340"} err="failed to get container status \"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\": rpc error: code = NotFound desc = could not find container \"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\": container with ID starting with e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.229400 4690 scope.go:117] "RemoveContainer" containerID="dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.229616 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\": container with ID starting with dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624 not found: ID does not exist" containerID="dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.229654 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624"} err="failed to get container status \"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\": rpc error: code = NotFound desc = could not find container \"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\": container with ID starting with dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.229670 4690 scope.go:117] "RemoveContainer" containerID="a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.229937 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\": container with ID starting with a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616 not found: ID does not exist" containerID="a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.230033 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616"} err="failed to get container status \"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\": rpc error: code = NotFound desc = could not find container \"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\": container with ID starting with a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.230064 4690 scope.go:117] "RemoveContainer" containerID="5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.230323 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\": container with ID starting with 5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8 not found: ID does not exist" containerID="5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.230350 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8"} err="failed to get container status \"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\": rpc error: code = NotFound desc = could not find container \"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\": container with ID starting with 5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.230388 4690 scope.go:117] "RemoveContainer" containerID="486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.230645 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\": container with ID starting with 486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b not found: ID does not exist" containerID="486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.230681 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b"} err="failed to get container status \"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\": rpc error: code = NotFound desc = could not find container \"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\": container with ID starting with 486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.230706 4690 scope.go:117] "RemoveContainer" containerID="515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.230944 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\": container with ID starting with 515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4 not found: ID does not exist" containerID="515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.230962 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4"} err="failed to get container status \"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\": rpc error: code = NotFound desc = could not find container \"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\": container with ID starting with 515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.230975 4690 scope.go:117] "RemoveContainer" containerID="deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.231193 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\": container with ID starting with deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf not found: ID does not exist" containerID="deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.231211 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf"} err="failed to get container status \"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\": rpc error: code = NotFound desc = could not find container \"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\": container with ID starting with deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.231227 4690 scope.go:117] "RemoveContainer" containerID="a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.231485 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\": container with ID starting with a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869 not found: ID does not exist" containerID="a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.231504 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869"} err="failed to get container status \"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\": rpc error: code = NotFound desc = could not find container \"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\": container with ID starting with a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.231516 4690 scope.go:117] "RemoveContainer" containerID="1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9" Feb 17 00:13:55 crc kubenswrapper[4690]: E0217 00:13:55.231725 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\": container with ID starting with 1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9 not found: ID does not exist" containerID="1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.231741 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9"} err="failed to get container status \"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\": rpc error: code = NotFound desc = could not find container \"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\": container with ID starting with 1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.231752 4690 scope.go:117] "RemoveContainer" containerID="e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.232045 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65"} err="failed to get container status \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": rpc error: code = NotFound desc = could not find container \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": container with ID starting with e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.232065 4690 scope.go:117] "RemoveContainer" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.232338 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340"} err="failed to get container status \"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\": rpc error: code = NotFound desc = could not find container \"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\": container with ID starting with e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.232403 4690 scope.go:117] "RemoveContainer" containerID="dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.232666 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624"} err="failed to get container status \"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\": rpc error: code = NotFound desc = could not find container \"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\": container with ID starting with dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.232686 4690 scope.go:117] "RemoveContainer" containerID="a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.233235 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616"} err="failed to get container status \"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\": rpc error: code = NotFound desc = could not find container \"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\": container with ID starting with a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.233307 4690 scope.go:117] "RemoveContainer" containerID="5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.233745 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8"} err="failed to get container status \"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\": rpc error: code = NotFound desc = could not find container \"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\": container with ID starting with 5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.233766 4690 scope.go:117] "RemoveContainer" containerID="486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.233956 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b"} err="failed to get container status \"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\": rpc error: code = NotFound desc = could not find container \"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\": container with ID starting with 486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.233968 4690 scope.go:117] "RemoveContainer" containerID="515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.234172 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4"} err="failed to get container status \"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\": rpc error: code = NotFound desc = could not find container \"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\": container with ID starting with 515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.234196 4690 scope.go:117] "RemoveContainer" containerID="deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.234409 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf"} err="failed to get container status \"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\": rpc error: code = NotFound desc = could not find container \"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\": container with ID starting with deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.234428 4690 scope.go:117] "RemoveContainer" containerID="a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.234628 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869"} err="failed to get container status \"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\": rpc error: code = NotFound desc = could not find container \"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\": container with ID starting with a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.234647 4690 scope.go:117] "RemoveContainer" containerID="1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.234848 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9"} err="failed to get container status \"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\": rpc error: code = NotFound desc = could not find container \"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\": container with ID starting with 1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.234869 4690 scope.go:117] "RemoveContainer" containerID="e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.235073 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65"} err="failed to get container status \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": rpc error: code = NotFound desc = could not find container \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": container with ID starting with e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.235092 4690 scope.go:117] "RemoveContainer" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.235333 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340"} err="failed to get container status \"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\": rpc error: code = NotFound desc = could not find container \"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\": container with ID starting with e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.235357 4690 scope.go:117] "RemoveContainer" containerID="dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.235584 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624"} err="failed to get container status \"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\": rpc error: code = NotFound desc = could not find container \"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\": container with ID starting with dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.235601 4690 scope.go:117] "RemoveContainer" containerID="a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.235782 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616"} err="failed to get container status \"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\": rpc error: code = NotFound desc = could not find container \"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\": container with ID starting with a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.235804 4690 scope.go:117] "RemoveContainer" containerID="5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.236053 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8"} err="failed to get container status \"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\": rpc error: code = NotFound desc = could not find container \"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\": container with ID starting with 5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.236080 4690 scope.go:117] "RemoveContainer" containerID="486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.236309 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b"} err="failed to get container status \"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\": rpc error: code = NotFound desc = could not find container \"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\": container with ID starting with 486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.236338 4690 scope.go:117] "RemoveContainer" containerID="515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.236576 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4"} err="failed to get container status \"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\": rpc error: code = NotFound desc = could not find container \"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\": container with ID starting with 515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.236593 4690 scope.go:117] "RemoveContainer" containerID="deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.236802 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf"} err="failed to get container status \"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\": rpc error: code = NotFound desc = could not find container \"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\": container with ID starting with deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.236830 4690 scope.go:117] "RemoveContainer" containerID="a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.237045 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869"} err="failed to get container status \"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\": rpc error: code = NotFound desc = could not find container \"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\": container with ID starting with a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.237064 4690 scope.go:117] "RemoveContainer" containerID="1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.237284 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9"} err="failed to get container status \"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\": rpc error: code = NotFound desc = could not find container \"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\": container with ID starting with 1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.237307 4690 scope.go:117] "RemoveContainer" containerID="e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.237537 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65"} err="failed to get container status \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": rpc error: code = NotFound desc = could not find container \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": container with ID starting with e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.237557 4690 scope.go:117] "RemoveContainer" containerID="e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.237776 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340"} err="failed to get container status \"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\": rpc error: code = NotFound desc = could not find container \"e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340\": container with ID starting with e93c278381031971bb69f6d00c8c356e55de0786ed9109ab6db0e42eccd51340 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.237796 4690 scope.go:117] "RemoveContainer" containerID="dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.238024 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624"} err="failed to get container status \"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\": rpc error: code = NotFound desc = could not find container \"dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624\": container with ID starting with dc0e9a75b3d679c0e3b28d0cfad146c7dbee69af38ace64579b10ce74ce86624 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.238047 4690 scope.go:117] "RemoveContainer" containerID="a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.238264 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616"} err="failed to get container status \"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\": rpc error: code = NotFound desc = could not find container \"a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616\": container with ID starting with a41f212873945179ddc92084c0e833f5102353d1643e547c30b9fa30f164a616 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.238281 4690 scope.go:117] "RemoveContainer" containerID="5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.238506 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8"} err="failed to get container status \"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\": rpc error: code = NotFound desc = could not find container \"5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8\": container with ID starting with 5ca785c180dfa73fd7be6d23fa09855cfdc86de0553d3154efacd58ca7f6fec8 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.238533 4690 scope.go:117] "RemoveContainer" containerID="486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.238744 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b"} err="failed to get container status \"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\": rpc error: code = NotFound desc = could not find container \"486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b\": container with ID starting with 486c6f02c4892fcf47d4b218eba37eed9c28b368d27e06105cbd51004b664d1b not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.238784 4690 scope.go:117] "RemoveContainer" containerID="515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.238958 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4"} err="failed to get container status \"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\": rpc error: code = NotFound desc = could not find container \"515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4\": container with ID starting with 515ab48e849c451fb932914b1f2f7396158b9040c8634898e8b0e15221bafbb4 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.239021 4690 scope.go:117] "RemoveContainer" containerID="deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.239229 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf"} err="failed to get container status \"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\": rpc error: code = NotFound desc = could not find container \"deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf\": container with ID starting with deae6c106f62ddece357189ff43970ccfd091834ff37c6b9ff7c5d19474dcebf not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.239245 4690 scope.go:117] "RemoveContainer" containerID="a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.239470 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869"} err="failed to get container status \"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\": rpc error: code = NotFound desc = could not find container \"a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869\": container with ID starting with a4d72a82144fb7619c65cce156b4d6a40a2085bc8a2dd3f6eec915137713c869 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.239493 4690 scope.go:117] "RemoveContainer" containerID="1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.239707 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9"} err="failed to get container status \"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\": rpc error: code = NotFound desc = could not find container \"1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9\": container with ID starting with 1f7569aec27dd91b14890a2e33058d9b722c367b02c6cad324808bb75fe3f7d9 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.239728 4690 scope.go:117] "RemoveContainer" containerID="e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.239948 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65"} err="failed to get container status \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": rpc error: code = NotFound desc = could not find container \"e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65\": container with ID starting with e308b4df60124e2e98d9fc08857b2f1cdedf7f1f04a15fea9b1bb25cbe444d65 not found: ID does not exist" Feb 17 00:13:55 crc kubenswrapper[4690]: I0217 00:13:55.254333 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:13:56 crc kubenswrapper[4690]: I0217 00:13:56.065714 4690 generic.go:334] "Generic (PLEG): container finished" podID="1332a26f-1afd-4af0-bde4-77186a1350b5" containerID="5cb3cd94a6b7cc45ded18331aeb48c61035c89ebd746d12c1868883d7456a094" exitCode=0 Feb 17 00:13:56 crc kubenswrapper[4690]: I0217 00:13:56.065822 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" event={"ID":"1332a26f-1afd-4af0-bde4-77186a1350b5","Type":"ContainerDied","Data":"5cb3cd94a6b7cc45ded18331aeb48c61035c89ebd746d12c1868883d7456a094"} Feb 17 00:13:56 crc kubenswrapper[4690]: I0217 00:13:56.066195 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" event={"ID":"1332a26f-1afd-4af0-bde4-77186a1350b5","Type":"ContainerStarted","Data":"b300515d515ad09a7fe87bea023d840020e92fa2d70a9c0782a0872b6538e4ab"} Feb 17 00:13:56 crc kubenswrapper[4690]: I0217 00:13:56.326067 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:13:56 crc kubenswrapper[4690]: I0217 00:13:56.326502 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:13:56 crc kubenswrapper[4690]: I0217 00:13:56.986959 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6eeb20f-1ee2-4d96-a53b-da30d60213f3" path="/var/lib/kubelet/pods/c6eeb20f-1ee2-4d96-a53b-da30d60213f3/volumes" Feb 17 00:13:57 crc kubenswrapper[4690]: I0217 00:13:57.073767 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" event={"ID":"1332a26f-1afd-4af0-bde4-77186a1350b5","Type":"ContainerStarted","Data":"6ebe778e5acfe106df855351a160e20d5f23d2f7ecbfa565fe10837022c3086b"} Feb 17 00:13:57 crc kubenswrapper[4690]: I0217 00:13:57.073811 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" event={"ID":"1332a26f-1afd-4af0-bde4-77186a1350b5","Type":"ContainerStarted","Data":"0731047b9cf03cd285c7868afc2a2bf74ef25f9050789c9ec174009755b137b4"} Feb 17 00:13:57 crc kubenswrapper[4690]: I0217 00:13:57.073823 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" event={"ID":"1332a26f-1afd-4af0-bde4-77186a1350b5","Type":"ContainerStarted","Data":"e1dae9a4f8fcbaf8cc112e6f3dd7bbec5f83abe4d8566b33aa6a586900d33089"} Feb 17 00:13:57 crc kubenswrapper[4690]: I0217 00:13:57.073836 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" event={"ID":"1332a26f-1afd-4af0-bde4-77186a1350b5","Type":"ContainerStarted","Data":"ec46d24db86e9e4618bea0518f43b449fc6015f5e3933817579b840f92caf050"} Feb 17 00:13:57 crc kubenswrapper[4690]: I0217 00:13:57.073847 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" event={"ID":"1332a26f-1afd-4af0-bde4-77186a1350b5","Type":"ContainerStarted","Data":"a4be801689e22840d012feb733b825c1f5c6dabccd082478b9f5ee185ceb282a"} Feb 17 00:13:57 crc kubenswrapper[4690]: I0217 00:13:57.073891 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" event={"ID":"1332a26f-1afd-4af0-bde4-77186a1350b5","Type":"ContainerStarted","Data":"af649d1bcd59885d83a617e23d1a7b382d861f1ae21585efc5518bfce3fe621d"} Feb 17 00:14:00 crc kubenswrapper[4690]: I0217 00:14:00.103927 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" event={"ID":"1332a26f-1afd-4af0-bde4-77186a1350b5","Type":"ContainerStarted","Data":"da5bfc1423214a9bd8e09b9aa8606ed1b79341c13b21ff411644d615ff26e750"} Feb 17 00:14:02 crc kubenswrapper[4690]: I0217 00:14:02.118981 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" event={"ID":"1332a26f-1afd-4af0-bde4-77186a1350b5","Type":"ContainerStarted","Data":"ab891d5dd382b21162765024bd96f6ea3bf91a6366ca534316f437176dfd6220"} Feb 17 00:14:02 crc kubenswrapper[4690]: I0217 00:14:02.119422 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:14:02 crc kubenswrapper[4690]: I0217 00:14:02.119439 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:14:02 crc kubenswrapper[4690]: I0217 00:14:02.119453 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:14:02 crc kubenswrapper[4690]: I0217 00:14:02.146018 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:14:02 crc kubenswrapper[4690]: I0217 00:14:02.148932 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:14:02 crc kubenswrapper[4690]: I0217 00:14:02.152457 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" podStartSLOduration=8.152440784 podStartE2EDuration="8.152440784s" podCreationTimestamp="2026-02-17 00:13:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:14:02.151122107 +0000 UTC m=+515.886834858" watchObservedRunningTime="2026-02-17 00:14:02.152440784 +0000 UTC m=+515.888153535" Feb 17 00:14:07 crc kubenswrapper[4690]: I0217 00:14:07.976311 4690 scope.go:117] "RemoveContainer" containerID="b9405af1a1a3ca87c5a1110233884fa8b8dad590f816aa1ac3af9d2e5dd32ff6" Feb 17 00:14:07 crc kubenswrapper[4690]: E0217 00:14:07.977006 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vb4v5_openshift-multus(797b14df-f7d0-419b-95f8-f02f25409e66)\"" pod="openshift-multus/multus-vb4v5" podUID="797b14df-f7d0-419b-95f8-f02f25409e66" Feb 17 00:14:20 crc kubenswrapper[4690]: I0217 00:14:20.976447 4690 scope.go:117] "RemoveContainer" containerID="b9405af1a1a3ca87c5a1110233884fa8b8dad590f816aa1ac3af9d2e5dd32ff6" Feb 17 00:14:21 crc kubenswrapper[4690]: I0217 00:14:21.245264 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/2.log" Feb 17 00:14:21 crc kubenswrapper[4690]: I0217 00:14:21.246427 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/1.log" Feb 17 00:14:21 crc kubenswrapper[4690]: I0217 00:14:21.246509 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vb4v5" event={"ID":"797b14df-f7d0-419b-95f8-f02f25409e66","Type":"ContainerStarted","Data":"35b4be57a5e6522a96c125420f3fa76f4c474b69390619958c9b17254d74a42e"} Feb 17 00:14:25 crc kubenswrapper[4690]: I0217 00:14:25.294846 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lmrvp" Feb 17 00:14:26 crc kubenswrapper[4690]: I0217 00:14:26.325841 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:14:26 crc kubenswrapper[4690]: I0217 00:14:26.325924 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:14:27 crc kubenswrapper[4690]: I0217 00:14:27.313108 4690 scope.go:117] "RemoveContainer" containerID="543db4edc9c9f0665faadcaa698cc5bebce64b5091f3c2697f530fdca61697ae" Feb 17 00:14:28 crc kubenswrapper[4690]: I0217 00:14:28.303777 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vb4v5_797b14df-f7d0-419b-95f8-f02f25409e66/kube-multus/2.log" Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.119625 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqtk7"] Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.120775 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fqtk7" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerName="registry-server" containerID="cri-o://9653deb103ebbb851ce804a8febe6d315cf14b53883caf69688abf4682840234" gracePeriod=30 Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.446518 4690 generic.go:334] "Generic (PLEG): container finished" podID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerID="9653deb103ebbb851ce804a8febe6d315cf14b53883caf69688abf4682840234" exitCode=0 Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.446566 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqtk7" event={"ID":"8151c43f-c4a2-4b6e-856d-298a43bd2f17","Type":"ContainerDied","Data":"9653deb103ebbb851ce804a8febe6d315cf14b53883caf69688abf4682840234"} Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.599992 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.689022 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-utilities\") pod \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.689475 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmdz6\" (UniqueName: \"kubernetes.io/projected/8151c43f-c4a2-4b6e-856d-298a43bd2f17-kube-api-access-nmdz6\") pod \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.689520 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-catalog-content\") pod \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\" (UID: \"8151c43f-c4a2-4b6e-856d-298a43bd2f17\") " Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.690255 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-utilities" (OuterVolumeSpecName: "utilities") pod "8151c43f-c4a2-4b6e-856d-298a43bd2f17" (UID: "8151c43f-c4a2-4b6e-856d-298a43bd2f17"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.701859 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8151c43f-c4a2-4b6e-856d-298a43bd2f17-kube-api-access-nmdz6" (OuterVolumeSpecName: "kube-api-access-nmdz6") pod "8151c43f-c4a2-4b6e-856d-298a43bd2f17" (UID: "8151c43f-c4a2-4b6e-856d-298a43bd2f17"). InnerVolumeSpecName "kube-api-access-nmdz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.710705 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8151c43f-c4a2-4b6e-856d-298a43bd2f17" (UID: "8151c43f-c4a2-4b6e-856d-298a43bd2f17"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.791485 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmdz6\" (UniqueName: \"kubernetes.io/projected/8151c43f-c4a2-4b6e-856d-298a43bd2f17-kube-api-access-nmdz6\") on node \"crc\" DevicePath \"\"" Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.791588 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:14:49 crc kubenswrapper[4690]: I0217 00:14:49.791602 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8151c43f-c4a2-4b6e-856d-298a43bd2f17-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:14:50 crc kubenswrapper[4690]: I0217 00:14:50.454208 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqtk7" event={"ID":"8151c43f-c4a2-4b6e-856d-298a43bd2f17","Type":"ContainerDied","Data":"edc2658e000e8940e7072beae0ee0ccc2669c7d341ab1b7fe4fd862015102064"} Feb 17 00:14:50 crc kubenswrapper[4690]: I0217 00:14:50.454464 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqtk7" Feb 17 00:14:50 crc kubenswrapper[4690]: I0217 00:14:50.454809 4690 scope.go:117] "RemoveContainer" containerID="9653deb103ebbb851ce804a8febe6d315cf14b53883caf69688abf4682840234" Feb 17 00:14:50 crc kubenswrapper[4690]: I0217 00:14:50.471726 4690 scope.go:117] "RemoveContainer" containerID="94c9dd6933eef1de8e551b4403f67043b9607d8b45bf9fcb6519630a355b40b4" Feb 17 00:14:50 crc kubenswrapper[4690]: I0217 00:14:50.487807 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqtk7"] Feb 17 00:14:50 crc kubenswrapper[4690]: I0217 00:14:50.491427 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqtk7"] Feb 17 00:14:50 crc kubenswrapper[4690]: I0217 00:14:50.501062 4690 scope.go:117] "RemoveContainer" containerID="05a35ed956445475b96f91928f11420e1eda12c4f568a66eaf4ee29a73512741" Feb 17 00:14:50 crc kubenswrapper[4690]: I0217 00:14:50.983100 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" path="/var/lib/kubelet/pods/8151c43f-c4a2-4b6e-856d-298a43bd2f17/volumes" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.545682 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64"] Feb 17 00:14:53 crc kubenswrapper[4690]: E0217 00:14:53.546208 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerName="extract-content" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.546226 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerName="extract-content" Feb 17 00:14:53 crc kubenswrapper[4690]: E0217 00:14:53.546249 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerName="extract-utilities" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.546259 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerName="extract-utilities" Feb 17 00:14:53 crc kubenswrapper[4690]: E0217 00:14:53.546270 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerName="registry-server" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.546280 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerName="registry-server" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.546433 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="8151c43f-c4a2-4b6e-856d-298a43bd2f17" containerName="registry-server" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.547447 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.549580 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.555955 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64"] Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.639803 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rltvc\" (UniqueName: \"kubernetes.io/projected/584ec9e5-3927-451e-bd2c-5c5014425c24-kube-api-access-rltvc\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.639869 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.639920 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.741068 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.741162 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rltvc\" (UniqueName: \"kubernetes.io/projected/584ec9e5-3927-451e-bd2c-5c5014425c24-kube-api-access-rltvc\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.741190 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.741607 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.741617 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.777126 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rltvc\" (UniqueName: \"kubernetes.io/projected/584ec9e5-3927-451e-bd2c-5c5014425c24-kube-api-access-rltvc\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:53 crc kubenswrapper[4690]: I0217 00:14:53.874231 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:14:54 crc kubenswrapper[4690]: I0217 00:14:54.124059 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64"] Feb 17 00:14:54 crc kubenswrapper[4690]: I0217 00:14:54.481052 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" event={"ID":"584ec9e5-3927-451e-bd2c-5c5014425c24","Type":"ContainerStarted","Data":"348cb7a227d6647f4223e0f786ed7c1c45832d561931e5733a8672538191b35c"} Feb 17 00:14:54 crc kubenswrapper[4690]: I0217 00:14:54.481133 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" event={"ID":"584ec9e5-3927-451e-bd2c-5c5014425c24","Type":"ContainerStarted","Data":"dfcde97eddfbcac9a54606f8875b6a09ef279e1971847f510a1a21e72df4cdd3"} Feb 17 00:14:55 crc kubenswrapper[4690]: I0217 00:14:55.487582 4690 generic.go:334] "Generic (PLEG): container finished" podID="584ec9e5-3927-451e-bd2c-5c5014425c24" containerID="348cb7a227d6647f4223e0f786ed7c1c45832d561931e5733a8672538191b35c" exitCode=0 Feb 17 00:14:55 crc kubenswrapper[4690]: I0217 00:14:55.487620 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" event={"ID":"584ec9e5-3927-451e-bd2c-5c5014425c24","Type":"ContainerDied","Data":"348cb7a227d6647f4223e0f786ed7c1c45832d561931e5733a8672538191b35c"} Feb 17 00:14:55 crc kubenswrapper[4690]: I0217 00:14:55.490190 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 00:14:56 crc kubenswrapper[4690]: I0217 00:14:56.325703 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:14:56 crc kubenswrapper[4690]: I0217 00:14:56.326118 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:14:56 crc kubenswrapper[4690]: I0217 00:14:56.326195 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:14:56 crc kubenswrapper[4690]: I0217 00:14:56.327220 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2bfdb84a832fb3eb0bc54fa1a1fa0b2de03c4db1987487e9015bfa4b3b7cb1bb"} pod="openshift-machine-config-operator/machine-config-daemon-lszwm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 00:14:56 crc kubenswrapper[4690]: I0217 00:14:56.327340 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" containerID="cri-o://2bfdb84a832fb3eb0bc54fa1a1fa0b2de03c4db1987487e9015bfa4b3b7cb1bb" gracePeriod=600 Feb 17 00:14:56 crc kubenswrapper[4690]: I0217 00:14:56.500801 4690 generic.go:334] "Generic (PLEG): container finished" podID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerID="2bfdb84a832fb3eb0bc54fa1a1fa0b2de03c4db1987487e9015bfa4b3b7cb1bb" exitCode=0 Feb 17 00:14:56 crc kubenswrapper[4690]: I0217 00:14:56.501017 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerDied","Data":"2bfdb84a832fb3eb0bc54fa1a1fa0b2de03c4db1987487e9015bfa4b3b7cb1bb"} Feb 17 00:14:56 crc kubenswrapper[4690]: I0217 00:14:56.501489 4690 scope.go:117] "RemoveContainer" containerID="994b411b25ff62fc16205928af96a4f9d2be45f24e3023cf34d1411551126d69" Feb 17 00:14:57 crc kubenswrapper[4690]: I0217 00:14:57.508139 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerStarted","Data":"957be8da43ec9afba6c8ee251740bb0bf80c9b7194d697b255095f03095c6cae"} Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.125831 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl"] Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.127835 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.142022 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl"] Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.222286 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.222618 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjgb6\" (UniqueName: \"kubernetes.io/projected/73d09007-760b-47ba-938a-db3ec30b065f-kube-api-access-hjgb6\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.222673 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.324191 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjgb6\" (UniqueName: \"kubernetes.io/projected/73d09007-760b-47ba-938a-db3ec30b065f-kube-api-access-hjgb6\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.324294 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.324329 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.325017 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.325063 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.345813 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjgb6\" (UniqueName: \"kubernetes.io/projected/73d09007-760b-47ba-938a-db3ec30b065f-kube-api-access-hjgb6\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.454152 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.724935 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl"] Feb 17 00:14:59 crc kubenswrapper[4690]: W0217 00:14:59.734845 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73d09007_760b_47ba_938a_db3ec30b065f.slice/crio-395b5c280553cfb68e94bf84e1094e8c153ef04aae4336658c502c2e1187dbf7 WatchSource:0}: Error finding container 395b5c280553cfb68e94bf84e1094e8c153ef04aae4336658c502c2e1187dbf7: Status 404 returned error can't find the container with id 395b5c280553cfb68e94bf84e1094e8c153ef04aae4336658c502c2e1187dbf7 Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.930244 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8"] Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.931256 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.932944 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.933245 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jpxv\" (UniqueName: \"kubernetes.io/projected/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-kube-api-access-6jpxv\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.933448 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:14:59 crc kubenswrapper[4690]: I0217 00:14:59.941415 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8"] Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.034161 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.034205 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.034257 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jpxv\" (UniqueName: \"kubernetes.io/projected/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-kube-api-access-6jpxv\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.034806 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.035340 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.054412 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jpxv\" (UniqueName: \"kubernetes.io/projected/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-kube-api-access-6jpxv\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.156724 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9"] Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.157638 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.159965 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.161475 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.167900 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9"] Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.262003 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.337664 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b04841b8-e534-415b-b5b1-7ed3f0127366-secret-volume\") pod \"collect-profiles-29521455-mpgj9\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.337726 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh2tz\" (UniqueName: \"kubernetes.io/projected/b04841b8-e534-415b-b5b1-7ed3f0127366-kube-api-access-fh2tz\") pod \"collect-profiles-29521455-mpgj9\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.337792 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b04841b8-e534-415b-b5b1-7ed3f0127366-config-volume\") pod \"collect-profiles-29521455-mpgj9\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.439634 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b04841b8-e534-415b-b5b1-7ed3f0127366-secret-volume\") pod \"collect-profiles-29521455-mpgj9\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.440223 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh2tz\" (UniqueName: \"kubernetes.io/projected/b04841b8-e534-415b-b5b1-7ed3f0127366-kube-api-access-fh2tz\") pod \"collect-profiles-29521455-mpgj9\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.440304 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b04841b8-e534-415b-b5b1-7ed3f0127366-config-volume\") pod \"collect-profiles-29521455-mpgj9\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.441453 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b04841b8-e534-415b-b5b1-7ed3f0127366-config-volume\") pod \"collect-profiles-29521455-mpgj9\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.456731 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b04841b8-e534-415b-b5b1-7ed3f0127366-secret-volume\") pod \"collect-profiles-29521455-mpgj9\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.462085 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh2tz\" (UniqueName: \"kubernetes.io/projected/b04841b8-e534-415b-b5b1-7ed3f0127366-kube-api-access-fh2tz\") pod \"collect-profiles-29521455-mpgj9\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.462890 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8"] Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.493482 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:00 crc kubenswrapper[4690]: W0217 00:15:00.516711 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c7e787f_4fb9_40ca_8ea8_331388ab61b0.slice/crio-842ac60573560923a834879604f15243d76cc6ec57695518393e00b2e8885c96 WatchSource:0}: Error finding container 842ac60573560923a834879604f15243d76cc6ec57695518393e00b2e8885c96: Status 404 returned error can't find the container with id 842ac60573560923a834879604f15243d76cc6ec57695518393e00b2e8885c96 Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.530843 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" event={"ID":"6c7e787f-4fb9-40ca-8ea8-331388ab61b0","Type":"ContainerStarted","Data":"842ac60573560923a834879604f15243d76cc6ec57695518393e00b2e8885c96"} Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.532678 4690 generic.go:334] "Generic (PLEG): container finished" podID="73d09007-760b-47ba-938a-db3ec30b065f" containerID="395b62661b085d9d6ae241c03ad83949b876f94367a3027fa9721593c4cb732f" exitCode=0 Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.532712 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" event={"ID":"73d09007-760b-47ba-938a-db3ec30b065f","Type":"ContainerDied","Data":"395b62661b085d9d6ae241c03ad83949b876f94367a3027fa9721593c4cb732f"} Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.532732 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" event={"ID":"73d09007-760b-47ba-938a-db3ec30b065f","Type":"ContainerStarted","Data":"395b5c280553cfb68e94bf84e1094e8c153ef04aae4336658c502c2e1187dbf7"} Feb 17 00:15:00 crc kubenswrapper[4690]: I0217 00:15:00.732966 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9"] Feb 17 00:15:00 crc kubenswrapper[4690]: W0217 00:15:00.743080 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb04841b8_e534_415b_b5b1_7ed3f0127366.slice/crio-514b4106e152dc2f5481ce57e8589e280cb9e8f40e1b1fa0ae5e56c136043e4c WatchSource:0}: Error finding container 514b4106e152dc2f5481ce57e8589e280cb9e8f40e1b1fa0ae5e56c136043e4c: Status 404 returned error can't find the container with id 514b4106e152dc2f5481ce57e8589e280cb9e8f40e1b1fa0ae5e56c136043e4c Feb 17 00:15:01 crc kubenswrapper[4690]: I0217 00:15:01.540523 4690 generic.go:334] "Generic (PLEG): container finished" podID="584ec9e5-3927-451e-bd2c-5c5014425c24" containerID="098a340bdfd5c9400e26120f2d79d5b3140b09ec6497cd42de20c12efd6f839a" exitCode=0 Feb 17 00:15:01 crc kubenswrapper[4690]: I0217 00:15:01.540653 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" event={"ID":"584ec9e5-3927-451e-bd2c-5c5014425c24","Type":"ContainerDied","Data":"098a340bdfd5c9400e26120f2d79d5b3140b09ec6497cd42de20c12efd6f839a"} Feb 17 00:15:01 crc kubenswrapper[4690]: I0217 00:15:01.545250 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" event={"ID":"6c7e787f-4fb9-40ca-8ea8-331388ab61b0","Type":"ContainerDied","Data":"09bac06c23e827ea213c0061c39c101b8e18409b5ac4b39ee47cc871ad241401"} Feb 17 00:15:01 crc kubenswrapper[4690]: I0217 00:15:01.545047 4690 generic.go:334] "Generic (PLEG): container finished" podID="6c7e787f-4fb9-40ca-8ea8-331388ab61b0" containerID="09bac06c23e827ea213c0061c39c101b8e18409b5ac4b39ee47cc871ad241401" exitCode=0 Feb 17 00:15:01 crc kubenswrapper[4690]: I0217 00:15:01.549101 4690 generic.go:334] "Generic (PLEG): container finished" podID="b04841b8-e534-415b-b5b1-7ed3f0127366" containerID="988941d9b14b52fc9797edb3319f4e7470cd898ba86d6282e70f7612cd68e496" exitCode=0 Feb 17 00:15:01 crc kubenswrapper[4690]: I0217 00:15:01.549203 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" event={"ID":"b04841b8-e534-415b-b5b1-7ed3f0127366","Type":"ContainerDied","Data":"988941d9b14b52fc9797edb3319f4e7470cd898ba86d6282e70f7612cd68e496"} Feb 17 00:15:01 crc kubenswrapper[4690]: I0217 00:15:01.551689 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" event={"ID":"b04841b8-e534-415b-b5b1-7ed3f0127366","Type":"ContainerStarted","Data":"514b4106e152dc2f5481ce57e8589e280cb9e8f40e1b1fa0ae5e56c136043e4c"} Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.556234 4690 generic.go:334] "Generic (PLEG): container finished" podID="584ec9e5-3927-451e-bd2c-5c5014425c24" containerID="deb8cba4232f8e289555ffdfcec727cce2f26fca30fc149541efa5aabb6a60ef" exitCode=0 Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.556451 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" event={"ID":"584ec9e5-3927-451e-bd2c-5c5014425c24","Type":"ContainerDied","Data":"deb8cba4232f8e289555ffdfcec727cce2f26fca30fc149541efa5aabb6a60ef"} Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.829931 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.871690 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh2tz\" (UniqueName: \"kubernetes.io/projected/b04841b8-e534-415b-b5b1-7ed3f0127366-kube-api-access-fh2tz\") pod \"b04841b8-e534-415b-b5b1-7ed3f0127366\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.871823 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b04841b8-e534-415b-b5b1-7ed3f0127366-secret-volume\") pod \"b04841b8-e534-415b-b5b1-7ed3f0127366\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.871869 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b04841b8-e534-415b-b5b1-7ed3f0127366-config-volume\") pod \"b04841b8-e534-415b-b5b1-7ed3f0127366\" (UID: \"b04841b8-e534-415b-b5b1-7ed3f0127366\") " Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.872927 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b04841b8-e534-415b-b5b1-7ed3f0127366-config-volume" (OuterVolumeSpecName: "config-volume") pod "b04841b8-e534-415b-b5b1-7ed3f0127366" (UID: "b04841b8-e534-415b-b5b1-7ed3f0127366"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.877492 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b04841b8-e534-415b-b5b1-7ed3f0127366-kube-api-access-fh2tz" (OuterVolumeSpecName: "kube-api-access-fh2tz") pod "b04841b8-e534-415b-b5b1-7ed3f0127366" (UID: "b04841b8-e534-415b-b5b1-7ed3f0127366"). InnerVolumeSpecName "kube-api-access-fh2tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.877533 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b04841b8-e534-415b-b5b1-7ed3f0127366-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b04841b8-e534-415b-b5b1-7ed3f0127366" (UID: "b04841b8-e534-415b-b5b1-7ed3f0127366"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.973338 4690 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b04841b8-e534-415b-b5b1-7ed3f0127366-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.973448 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b04841b8-e534-415b-b5b1-7ed3f0127366-config-volume\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:02 crc kubenswrapper[4690]: I0217 00:15:02.973464 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh2tz\" (UniqueName: \"kubernetes.io/projected/b04841b8-e534-415b-b5b1-7ed3f0127366-kube-api-access-fh2tz\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:03 crc kubenswrapper[4690]: I0217 00:15:03.568137 4690 generic.go:334] "Generic (PLEG): container finished" podID="73d09007-760b-47ba-938a-db3ec30b065f" containerID="c0b15032404ae3bbc68c21cf4b3331f84291da63fbe2a2524dc7592051876220" exitCode=0 Feb 17 00:15:03 crc kubenswrapper[4690]: I0217 00:15:03.568347 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" event={"ID":"73d09007-760b-47ba-938a-db3ec30b065f","Type":"ContainerDied","Data":"c0b15032404ae3bbc68c21cf4b3331f84291da63fbe2a2524dc7592051876220"} Feb 17 00:15:03 crc kubenswrapper[4690]: I0217 00:15:03.571063 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" Feb 17 00:15:03 crc kubenswrapper[4690]: I0217 00:15:03.571142 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29521455-mpgj9" event={"ID":"b04841b8-e534-415b-b5b1-7ed3f0127366","Type":"ContainerDied","Data":"514b4106e152dc2f5481ce57e8589e280cb9e8f40e1b1fa0ae5e56c136043e4c"} Feb 17 00:15:03 crc kubenswrapper[4690]: I0217 00:15:03.571187 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="514b4106e152dc2f5481ce57e8589e280cb9e8f40e1b1fa0ae5e56c136043e4c" Feb 17 00:15:03 crc kubenswrapper[4690]: I0217 00:15:03.953294 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.129444 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-bundle\") pod \"584ec9e5-3927-451e-bd2c-5c5014425c24\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.129540 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-util\") pod \"584ec9e5-3927-451e-bd2c-5c5014425c24\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.129603 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rltvc\" (UniqueName: \"kubernetes.io/projected/584ec9e5-3927-451e-bd2c-5c5014425c24-kube-api-access-rltvc\") pod \"584ec9e5-3927-451e-bd2c-5c5014425c24\" (UID: \"584ec9e5-3927-451e-bd2c-5c5014425c24\") " Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.137880 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-bundle" (OuterVolumeSpecName: "bundle") pod "584ec9e5-3927-451e-bd2c-5c5014425c24" (UID: "584ec9e5-3927-451e-bd2c-5c5014425c24"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.138498 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/584ec9e5-3927-451e-bd2c-5c5014425c24-kube-api-access-rltvc" (OuterVolumeSpecName: "kube-api-access-rltvc") pod "584ec9e5-3927-451e-bd2c-5c5014425c24" (UID: "584ec9e5-3927-451e-bd2c-5c5014425c24"). InnerVolumeSpecName "kube-api-access-rltvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.144379 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-util" (OuterVolumeSpecName: "util") pod "584ec9e5-3927-451e-bd2c-5c5014425c24" (UID: "584ec9e5-3927-451e-bd2c-5c5014425c24"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.230738 4690 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.230772 4690 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/584ec9e5-3927-451e-bd2c-5c5014425c24-util\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.230782 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rltvc\" (UniqueName: \"kubernetes.io/projected/584ec9e5-3927-451e-bd2c-5c5014425c24-kube-api-access-rltvc\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.579242 4690 generic.go:334] "Generic (PLEG): container finished" podID="6c7e787f-4fb9-40ca-8ea8-331388ab61b0" containerID="793d77570171b44e8672a78676b42f797e873626852c8f6f7e30c9bff548d75b" exitCode=0 Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.579404 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" event={"ID":"6c7e787f-4fb9-40ca-8ea8-331388ab61b0","Type":"ContainerDied","Data":"793d77570171b44e8672a78676b42f797e873626852c8f6f7e30c9bff548d75b"} Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.582071 4690 generic.go:334] "Generic (PLEG): container finished" podID="73d09007-760b-47ba-938a-db3ec30b065f" containerID="8116dd1c8d240455e811aacc3d0fe72e9603cb57b06844c7aad284b405f1c0f5" exitCode=0 Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.582213 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" event={"ID":"73d09007-760b-47ba-938a-db3ec30b065f","Type":"ContainerDied","Data":"8116dd1c8d240455e811aacc3d0fe72e9603cb57b06844c7aad284b405f1c0f5"} Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.590510 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" event={"ID":"584ec9e5-3927-451e-bd2c-5c5014425c24","Type":"ContainerDied","Data":"dfcde97eddfbcac9a54606f8875b6a09ef279e1971847f510a1a21e72df4cdd3"} Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.590585 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfcde97eddfbcac9a54606f8875b6a09ef279e1971847f510a1a21e72df4cdd3" Feb 17 00:15:04 crc kubenswrapper[4690]: I0217 00:15:04.590864 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.599976 4690 generic.go:334] "Generic (PLEG): container finished" podID="6c7e787f-4fb9-40ca-8ea8-331388ab61b0" containerID="1d558dc7a3d4cfc14ae585862697ee5bffa0aed847cf296b53530a113cd0bd17" exitCode=0 Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.600063 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" event={"ID":"6c7e787f-4fb9-40ca-8ea8-331388ab61b0","Type":"ContainerDied","Data":"1d558dc7a3d4cfc14ae585862697ee5bffa0aed847cf296b53530a113cd0bd17"} Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.856723 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.921315 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c"] Feb 17 00:15:05 crc kubenswrapper[4690]: E0217 00:15:05.921602 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73d09007-760b-47ba-938a-db3ec30b065f" containerName="extract" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.921622 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="73d09007-760b-47ba-938a-db3ec30b065f" containerName="extract" Feb 17 00:15:05 crc kubenswrapper[4690]: E0217 00:15:05.921643 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="584ec9e5-3927-451e-bd2c-5c5014425c24" containerName="extract" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.921657 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="584ec9e5-3927-451e-bd2c-5c5014425c24" containerName="extract" Feb 17 00:15:05 crc kubenswrapper[4690]: E0217 00:15:05.921679 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="584ec9e5-3927-451e-bd2c-5c5014425c24" containerName="util" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.921691 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="584ec9e5-3927-451e-bd2c-5c5014425c24" containerName="util" Feb 17 00:15:05 crc kubenswrapper[4690]: E0217 00:15:05.921705 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73d09007-760b-47ba-938a-db3ec30b065f" containerName="util" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.921717 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="73d09007-760b-47ba-938a-db3ec30b065f" containerName="util" Feb 17 00:15:05 crc kubenswrapper[4690]: E0217 00:15:05.921731 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="584ec9e5-3927-451e-bd2c-5c5014425c24" containerName="pull" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.921743 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="584ec9e5-3927-451e-bd2c-5c5014425c24" containerName="pull" Feb 17 00:15:05 crc kubenswrapper[4690]: E0217 00:15:05.921761 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b04841b8-e534-415b-b5b1-7ed3f0127366" containerName="collect-profiles" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.921776 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b04841b8-e534-415b-b5b1-7ed3f0127366" containerName="collect-profiles" Feb 17 00:15:05 crc kubenswrapper[4690]: E0217 00:15:05.921800 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73d09007-760b-47ba-938a-db3ec30b065f" containerName="pull" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.921812 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="73d09007-760b-47ba-938a-db3ec30b065f" containerName="pull" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.921980 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="b04841b8-e534-415b-b5b1-7ed3f0127366" containerName="collect-profiles" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.922026 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="584ec9e5-3927-451e-bd2c-5c5014425c24" containerName="extract" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.922060 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="73d09007-760b-47ba-938a-db3ec30b065f" containerName="extract" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.923343 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:05 crc kubenswrapper[4690]: I0217 00:15:05.930230 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c"] Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.052030 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-util\") pod \"73d09007-760b-47ba-938a-db3ec30b065f\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.052142 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-bundle\") pod \"73d09007-760b-47ba-938a-db3ec30b065f\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.052205 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjgb6\" (UniqueName: \"kubernetes.io/projected/73d09007-760b-47ba-938a-db3ec30b065f-kube-api-access-hjgb6\") pod \"73d09007-760b-47ba-938a-db3ec30b065f\" (UID: \"73d09007-760b-47ba-938a-db3ec30b065f\") " Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.052495 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwkxc\" (UniqueName: \"kubernetes.io/projected/930903ed-43cf-4146-bf5a-56f6f9d16434-kube-api-access-dwkxc\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.052591 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.052627 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.052777 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-bundle" (OuterVolumeSpecName: "bundle") pod "73d09007-760b-47ba-938a-db3ec30b065f" (UID: "73d09007-760b-47ba-938a-db3ec30b065f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.062656 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73d09007-760b-47ba-938a-db3ec30b065f-kube-api-access-hjgb6" (OuterVolumeSpecName: "kube-api-access-hjgb6") pod "73d09007-760b-47ba-938a-db3ec30b065f" (UID: "73d09007-760b-47ba-938a-db3ec30b065f"). InnerVolumeSpecName "kube-api-access-hjgb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.072063 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-util" (OuterVolumeSpecName: "util") pod "73d09007-760b-47ba-938a-db3ec30b065f" (UID: "73d09007-760b-47ba-938a-db3ec30b065f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.154141 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.154188 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.154298 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwkxc\" (UniqueName: \"kubernetes.io/projected/930903ed-43cf-4146-bf5a-56f6f9d16434-kube-api-access-dwkxc\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.154420 4690 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-util\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.154486 4690 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73d09007-760b-47ba-938a-db3ec30b065f-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.154654 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjgb6\" (UniqueName: \"kubernetes.io/projected/73d09007-760b-47ba-938a-db3ec30b065f-kube-api-access-hjgb6\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.154678 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.154876 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.172931 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwkxc\" (UniqueName: \"kubernetes.io/projected/930903ed-43cf-4146-bf5a-56f6f9d16434-kube-api-access-dwkxc\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.255343 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.612147 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.612134 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl" event={"ID":"73d09007-760b-47ba-938a-db3ec30b065f","Type":"ContainerDied","Data":"395b5c280553cfb68e94bf84e1094e8c153ef04aae4336658c502c2e1187dbf7"} Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.612678 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="395b5c280553cfb68e94bf84e1094e8c153ef04aae4336658c502c2e1187dbf7" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.684410 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c"] Feb 17 00:15:06 crc kubenswrapper[4690]: W0217 00:15:06.693207 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod930903ed_43cf_4146_bf5a_56f6f9d16434.slice/crio-c52879c0ec469fe241452b01abb0a3dc352a88cc885254f2f57a813f1176113c WatchSource:0}: Error finding container c52879c0ec469fe241452b01abb0a3dc352a88cc885254f2f57a813f1176113c: Status 404 returned error can't find the container with id c52879c0ec469fe241452b01abb0a3dc352a88cc885254f2f57a813f1176113c Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.886227 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.971929 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jpxv\" (UniqueName: \"kubernetes.io/projected/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-kube-api-access-6jpxv\") pod \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.971992 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-bundle\") pod \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.972035 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-util\") pod \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\" (UID: \"6c7e787f-4fb9-40ca-8ea8-331388ab61b0\") " Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.972817 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-bundle" (OuterVolumeSpecName: "bundle") pod "6c7e787f-4fb9-40ca-8ea8-331388ab61b0" (UID: "6c7e787f-4fb9-40ca-8ea8-331388ab61b0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.976331 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-kube-api-access-6jpxv" (OuterVolumeSpecName: "kube-api-access-6jpxv") pod "6c7e787f-4fb9-40ca-8ea8-331388ab61b0" (UID: "6c7e787f-4fb9-40ca-8ea8-331388ab61b0"). InnerVolumeSpecName "kube-api-access-6jpxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.988719 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-util" (OuterVolumeSpecName: "util") pod "6c7e787f-4fb9-40ca-8ea8-331388ab61b0" (UID: "6c7e787f-4fb9-40ca-8ea8-331388ab61b0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.989138 4690 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-util\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.989165 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jpxv\" (UniqueName: \"kubernetes.io/projected/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-kube-api-access-6jpxv\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:06 crc kubenswrapper[4690]: I0217 00:15:06.989179 4690 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c7e787f-4fb9-40ca-8ea8-331388ab61b0-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:07 crc kubenswrapper[4690]: I0217 00:15:07.620230 4690 generic.go:334] "Generic (PLEG): container finished" podID="930903ed-43cf-4146-bf5a-56f6f9d16434" containerID="b9b2c28e56e366335fe7d932ca4a799b8460489437e563f8c77cd2cf77f6dc07" exitCode=0 Feb 17 00:15:07 crc kubenswrapper[4690]: I0217 00:15:07.620323 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" event={"ID":"930903ed-43cf-4146-bf5a-56f6f9d16434","Type":"ContainerDied","Data":"b9b2c28e56e366335fe7d932ca4a799b8460489437e563f8c77cd2cf77f6dc07"} Feb 17 00:15:07 crc kubenswrapper[4690]: I0217 00:15:07.620443 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" event={"ID":"930903ed-43cf-4146-bf5a-56f6f9d16434","Type":"ContainerStarted","Data":"c52879c0ec469fe241452b01abb0a3dc352a88cc885254f2f57a813f1176113c"} Feb 17 00:15:07 crc kubenswrapper[4690]: I0217 00:15:07.627866 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" event={"ID":"6c7e787f-4fb9-40ca-8ea8-331388ab61b0","Type":"ContainerDied","Data":"842ac60573560923a834879604f15243d76cc6ec57695518393e00b2e8885c96"} Feb 17 00:15:07 crc kubenswrapper[4690]: I0217 00:15:07.627922 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="842ac60573560923a834879604f15243d76cc6ec57695518393e00b2e8885c96" Feb 17 00:15:07 crc kubenswrapper[4690]: I0217 00:15:07.627958 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8" Feb 17 00:15:11 crc kubenswrapper[4690]: I0217 00:15:11.667793 4690 generic.go:334] "Generic (PLEG): container finished" podID="930903ed-43cf-4146-bf5a-56f6f9d16434" containerID="9a6934e4692fddcbd0604bbf938f392ef510b5ce345adccfd2a0ec543a4fc075" exitCode=0 Feb 17 00:15:11 crc kubenswrapper[4690]: I0217 00:15:11.667904 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" event={"ID":"930903ed-43cf-4146-bf5a-56f6f9d16434","Type":"ContainerDied","Data":"9a6934e4692fddcbd0604bbf938f392ef510b5ce345adccfd2a0ec543a4fc075"} Feb 17 00:15:12 crc kubenswrapper[4690]: I0217 00:15:12.674906 4690 generic.go:334] "Generic (PLEG): container finished" podID="930903ed-43cf-4146-bf5a-56f6f9d16434" containerID="35d8037ffb004b07696b497b3f9dc3d0e75290d81f49d73498bb23cecdbcad0a" exitCode=0 Feb 17 00:15:12 crc kubenswrapper[4690]: I0217 00:15:12.674956 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" event={"ID":"930903ed-43cf-4146-bf5a-56f6f9d16434","Type":"ContainerDied","Data":"35d8037ffb004b07696b497b3f9dc3d0e75290d81f49d73498bb23cecdbcad0a"} Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.049772 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.095179 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-util\") pod \"930903ed-43cf-4146-bf5a-56f6f9d16434\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.095307 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwkxc\" (UniqueName: \"kubernetes.io/projected/930903ed-43cf-4146-bf5a-56f6f9d16434-kube-api-access-dwkxc\") pod \"930903ed-43cf-4146-bf5a-56f6f9d16434\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.095330 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-bundle\") pod \"930903ed-43cf-4146-bf5a-56f6f9d16434\" (UID: \"930903ed-43cf-4146-bf5a-56f6f9d16434\") " Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.096207 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-bundle" (OuterVolumeSpecName: "bundle") pod "930903ed-43cf-4146-bf5a-56f6f9d16434" (UID: "930903ed-43cf-4146-bf5a-56f6f9d16434"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.103523 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/930903ed-43cf-4146-bf5a-56f6f9d16434-kube-api-access-dwkxc" (OuterVolumeSpecName: "kube-api-access-dwkxc") pod "930903ed-43cf-4146-bf5a-56f6f9d16434" (UID: "930903ed-43cf-4146-bf5a-56f6f9d16434"). InnerVolumeSpecName "kube-api-access-dwkxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.106283 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-util" (OuterVolumeSpecName: "util") pod "930903ed-43cf-4146-bf5a-56f6f9d16434" (UID: "930903ed-43cf-4146-bf5a-56f6f9d16434"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.196386 4690 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-util\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.196415 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwkxc\" (UniqueName: \"kubernetes.io/projected/930903ed-43cf-4146-bf5a-56f6f9d16434-kube-api-access-dwkxc\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.196427 4690 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/930903ed-43cf-4146-bf5a-56f6f9d16434-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.685334 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" event={"ID":"930903ed-43cf-4146-bf5a-56f6f9d16434","Type":"ContainerDied","Data":"c52879c0ec469fe241452b01abb0a3dc352a88cc885254f2f57a813f1176113c"} Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.685391 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c52879c0ec469fe241452b01abb0a3dc352a88cc885254f2f57a813f1176113c" Feb 17 00:15:14 crc kubenswrapper[4690]: I0217 00:15:14.685413 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.365991 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-86q9q"] Feb 17 00:15:17 crc kubenswrapper[4690]: E0217 00:15:17.366498 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7e787f-4fb9-40ca-8ea8-331388ab61b0" containerName="pull" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.366513 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7e787f-4fb9-40ca-8ea8-331388ab61b0" containerName="pull" Feb 17 00:15:17 crc kubenswrapper[4690]: E0217 00:15:17.366524 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="930903ed-43cf-4146-bf5a-56f6f9d16434" containerName="util" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.366531 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="930903ed-43cf-4146-bf5a-56f6f9d16434" containerName="util" Feb 17 00:15:17 crc kubenswrapper[4690]: E0217 00:15:17.366542 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7e787f-4fb9-40ca-8ea8-331388ab61b0" containerName="extract" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.366549 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7e787f-4fb9-40ca-8ea8-331388ab61b0" containerName="extract" Feb 17 00:15:17 crc kubenswrapper[4690]: E0217 00:15:17.366561 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="930903ed-43cf-4146-bf5a-56f6f9d16434" containerName="extract" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.366568 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="930903ed-43cf-4146-bf5a-56f6f9d16434" containerName="extract" Feb 17 00:15:17 crc kubenswrapper[4690]: E0217 00:15:17.366584 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7e787f-4fb9-40ca-8ea8-331388ab61b0" containerName="util" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.366592 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7e787f-4fb9-40ca-8ea8-331388ab61b0" containerName="util" Feb 17 00:15:17 crc kubenswrapper[4690]: E0217 00:15:17.366599 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="930903ed-43cf-4146-bf5a-56f6f9d16434" containerName="pull" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.366605 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="930903ed-43cf-4146-bf5a-56f6f9d16434" containerName="pull" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.366710 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c7e787f-4fb9-40ca-8ea8-331388ab61b0" containerName="extract" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.366722 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="930903ed-43cf-4146-bf5a-56f6f9d16434" containerName="extract" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.367115 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-86q9q" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.369111 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-64smm" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.372934 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.375148 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.378295 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-86q9q"] Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.433755 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq5cb\" (UniqueName: \"kubernetes.io/projected/4e01fa77-1ce1-4285-9845-722ccd914395-kube-api-access-kq5cb\") pod \"interconnect-operator-5bb49f789d-86q9q\" (UID: \"4e01fa77-1ce1-4285-9845-722ccd914395\") " pod="service-telemetry/interconnect-operator-5bb49f789d-86q9q" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.535392 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq5cb\" (UniqueName: \"kubernetes.io/projected/4e01fa77-1ce1-4285-9845-722ccd914395-kube-api-access-kq5cb\") pod \"interconnect-operator-5bb49f789d-86q9q\" (UID: \"4e01fa77-1ce1-4285-9845-722ccd914395\") " pod="service-telemetry/interconnect-operator-5bb49f789d-86q9q" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.568459 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq5cb\" (UniqueName: \"kubernetes.io/projected/4e01fa77-1ce1-4285-9845-722ccd914395-kube-api-access-kq5cb\") pod \"interconnect-operator-5bb49f789d-86q9q\" (UID: \"4e01fa77-1ce1-4285-9845-722ccd914395\") " pod="service-telemetry/interconnect-operator-5bb49f789d-86q9q" Feb 17 00:15:17 crc kubenswrapper[4690]: I0217 00:15:17.683270 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-86q9q" Feb 17 00:15:18 crc kubenswrapper[4690]: I0217 00:15:18.526179 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-86q9q"] Feb 17 00:15:18 crc kubenswrapper[4690]: I0217 00:15:18.703640 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-86q9q" event={"ID":"4e01fa77-1ce1-4285-9845-722ccd914395","Type":"ContainerStarted","Data":"bc8d559a90dc937e987a456d165ea3fe16f023f9b4acf088bfe0c737d8f01c22"} Feb 17 00:15:21 crc kubenswrapper[4690]: I0217 00:15:21.904370 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb"] Feb 17 00:15:21 crc kubenswrapper[4690]: I0217 00:15:21.905153 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb" Feb 17 00:15:21 crc kubenswrapper[4690]: I0217 00:15:21.906925 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Feb 17 00:15:21 crc kubenswrapper[4690]: I0217 00:15:21.908274 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Feb 17 00:15:21 crc kubenswrapper[4690]: I0217 00:15:21.908274 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-wgdfj" Feb 17 00:15:21 crc kubenswrapper[4690]: I0217 00:15:21.921837 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb"] Feb 17 00:15:21 crc kubenswrapper[4690]: I0217 00:15:21.995145 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2fh2\" (UniqueName: \"kubernetes.io/projected/80f87711-62d4-488b-8a93-3dc6d93e3cc4-kube-api-access-l2fh2\") pod \"obo-prometheus-operator-68bc856cb9-hjdqb\" (UID: \"80f87711-62d4-488b-8a93-3dc6d93e3cc4\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.036233 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-69c756cfd-9qjz2"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.036892 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.039478 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.039854 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-t6cxx" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.062830 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.063446 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.066492 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.066798 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-hsgld" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.083994 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.084655 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.090828 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.096380 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4c6248a-697b-47eb-b7af-603a2bdf7ec0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg\" (UID: \"f4c6248a-697b-47eb-b7af-603a2bdf7ec0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.096432 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7-webhook-cert\") pod \"elastic-operator-69c756cfd-9qjz2\" (UID: \"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7\") " pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.096463 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2fh2\" (UniqueName: \"kubernetes.io/projected/80f87711-62d4-488b-8a93-3dc6d93e3cc4-kube-api-access-l2fh2\") pod \"obo-prometheus-operator-68bc856cb9-hjdqb\" (UID: \"80f87711-62d4-488b-8a93-3dc6d93e3cc4\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.096490 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4c6248a-697b-47eb-b7af-603a2bdf7ec0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg\" (UID: \"f4c6248a-697b-47eb-b7af-603a2bdf7ec0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.096523 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7-apiservice-cert\") pod \"elastic-operator-69c756cfd-9qjz2\" (UID: \"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7\") " pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.096550 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt92f\" (UniqueName: \"kubernetes.io/projected/5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7-kube-api-access-dt92f\") pod \"elastic-operator-69c756cfd-9qjz2\" (UID: \"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7\") " pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.107764 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.110571 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-69c756cfd-9qjz2"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.133408 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2fh2\" (UniqueName: \"kubernetes.io/projected/80f87711-62d4-488b-8a93-3dc6d93e3cc4-kube-api-access-l2fh2\") pod \"obo-prometheus-operator-68bc856cb9-hjdqb\" (UID: \"80f87711-62d4-488b-8a93-3dc6d93e3cc4\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.197486 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dab3686a-8cc4-4105-8460-a534c5d7558d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj\" (UID: \"dab3686a-8cc4-4105-8460-a534c5d7558d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.197538 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7-apiservice-cert\") pod \"elastic-operator-69c756cfd-9qjz2\" (UID: \"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7\") " pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.197576 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt92f\" (UniqueName: \"kubernetes.io/projected/5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7-kube-api-access-dt92f\") pod \"elastic-operator-69c756cfd-9qjz2\" (UID: \"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7\") " pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.197607 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dab3686a-8cc4-4105-8460-a534c5d7558d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj\" (UID: \"dab3686a-8cc4-4105-8460-a534c5d7558d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.197632 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4c6248a-697b-47eb-b7af-603a2bdf7ec0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg\" (UID: \"f4c6248a-697b-47eb-b7af-603a2bdf7ec0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.197652 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7-webhook-cert\") pod \"elastic-operator-69c756cfd-9qjz2\" (UID: \"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7\") " pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.197682 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4c6248a-697b-47eb-b7af-603a2bdf7ec0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg\" (UID: \"f4c6248a-697b-47eb-b7af-603a2bdf7ec0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.201670 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7-apiservice-cert\") pod \"elastic-operator-69c756cfd-9qjz2\" (UID: \"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7\") " pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.201921 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4c6248a-697b-47eb-b7af-603a2bdf7ec0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg\" (UID: \"f4c6248a-697b-47eb-b7af-603a2bdf7ec0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.205775 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4c6248a-697b-47eb-b7af-603a2bdf7ec0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg\" (UID: \"f4c6248a-697b-47eb-b7af-603a2bdf7ec0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.207217 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7-webhook-cert\") pod \"elastic-operator-69c756cfd-9qjz2\" (UID: \"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7\") " pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.216755 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt92f\" (UniqueName: \"kubernetes.io/projected/5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7-kube-api-access-dt92f\") pod \"elastic-operator-69c756cfd-9qjz2\" (UID: \"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7\") " pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.220651 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.251803 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-7mmxh"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.252480 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.254908 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-6r8wf" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.255592 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.268337 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-7mmxh"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.305951 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dab3686a-8cc4-4105-8460-a534c5d7558d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj\" (UID: \"dab3686a-8cc4-4105-8460-a534c5d7558d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.306009 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bae5780c-307b-4f82-8759-bde07af42405-observability-operator-tls\") pod \"observability-operator-59bdc8b94-7mmxh\" (UID: \"bae5780c-307b-4f82-8759-bde07af42405\") " pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.306028 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kcrd\" (UniqueName: \"kubernetes.io/projected/bae5780c-307b-4f82-8759-bde07af42405-kube-api-access-4kcrd\") pod \"observability-operator-59bdc8b94-7mmxh\" (UID: \"bae5780c-307b-4f82-8759-bde07af42405\") " pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.306060 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dab3686a-8cc4-4105-8460-a534c5d7558d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj\" (UID: \"dab3686a-8cc4-4105-8460-a534c5d7558d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.312159 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dab3686a-8cc4-4105-8460-a534c5d7558d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj\" (UID: \"dab3686a-8cc4-4105-8460-a534c5d7558d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.323633 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dab3686a-8cc4-4105-8460-a534c5d7558d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj\" (UID: \"dab3686a-8cc4-4105-8460-a534c5d7558d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.351016 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.381172 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.395892 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.407115 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bae5780c-307b-4f82-8759-bde07af42405-observability-operator-tls\") pod \"observability-operator-59bdc8b94-7mmxh\" (UID: \"bae5780c-307b-4f82-8759-bde07af42405\") " pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.407161 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kcrd\" (UniqueName: \"kubernetes.io/projected/bae5780c-307b-4f82-8759-bde07af42405-kube-api-access-4kcrd\") pod \"observability-operator-59bdc8b94-7mmxh\" (UID: \"bae5780c-307b-4f82-8759-bde07af42405\") " pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.418158 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bae5780c-307b-4f82-8759-bde07af42405-observability-operator-tls\") pod \"observability-operator-59bdc8b94-7mmxh\" (UID: \"bae5780c-307b-4f82-8759-bde07af42405\") " pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.423015 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kcrd\" (UniqueName: \"kubernetes.io/projected/bae5780c-307b-4f82-8759-bde07af42405-kube-api-access-4kcrd\") pod \"observability-operator-59bdc8b94-7mmxh\" (UID: \"bae5780c-307b-4f82-8759-bde07af42405\") " pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.452658 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-dj5w8"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.453347 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.455673 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-dtzmd" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.472332 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-dj5w8"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.508803 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d6e1fe6c-28db-455c-b460-827567af661d-openshift-service-ca\") pod \"perses-operator-5bf474d74f-dj5w8\" (UID: \"d6e1fe6c-28db-455c-b460-827567af661d\") " pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.508855 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjf25\" (UniqueName: \"kubernetes.io/projected/d6e1fe6c-28db-455c-b460-827567af661d-kube-api-access-xjf25\") pod \"perses-operator-5bf474d74f-dj5w8\" (UID: \"d6e1fe6c-28db-455c-b460-827567af661d\") " pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.595753 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.611994 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d6e1fe6c-28db-455c-b460-827567af661d-openshift-service-ca\") pod \"perses-operator-5bf474d74f-dj5w8\" (UID: \"d6e1fe6c-28db-455c-b460-827567af661d\") " pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.612052 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjf25\" (UniqueName: \"kubernetes.io/projected/d6e1fe6c-28db-455c-b460-827567af661d-kube-api-access-xjf25\") pod \"perses-operator-5bf474d74f-dj5w8\" (UID: \"d6e1fe6c-28db-455c-b460-827567af661d\") " pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.612938 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d6e1fe6c-28db-455c-b460-827567af661d-openshift-service-ca\") pod \"perses-operator-5bf474d74f-dj5w8\" (UID: \"d6e1fe6c-28db-455c-b460-827567af661d\") " pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.632087 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjf25\" (UniqueName: \"kubernetes.io/projected/d6e1fe6c-28db-455c-b460-827567af661d-kube-api-access-xjf25\") pod \"perses-operator-5bf474d74f-dj5w8\" (UID: \"d6e1fe6c-28db-455c-b460-827567af661d\") " pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.659082 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg"] Feb 17 00:15:22 crc kubenswrapper[4690]: W0217 00:15:22.660113 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4c6248a_697b_47eb_b7af_603a2bdf7ec0.slice/crio-1aa12cf5742f6cf9430e8ed4a583f458cc53a74c0fea9c9e169b215131122284 WatchSource:0}: Error finding container 1aa12cf5742f6cf9430e8ed4a583f458cc53a74c0fea9c9e169b215131122284: Status 404 returned error can't find the container with id 1aa12cf5742f6cf9430e8ed4a583f458cc53a74c0fea9c9e169b215131122284 Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.728582 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" event={"ID":"f4c6248a-697b-47eb-b7af-603a2bdf7ec0","Type":"ContainerStarted","Data":"1aa12cf5742f6cf9430e8ed4a583f458cc53a74c0fea9c9e169b215131122284"} Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.745487 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj"] Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.762748 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb"] Feb 17 00:15:22 crc kubenswrapper[4690]: W0217 00:15:22.781907 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddab3686a_8cc4_4105_8460_a534c5d7558d.slice/crio-42292750073adc527222839210572c295bbb02e412cfb0d636ea4c27ac35a3c4 WatchSource:0}: Error finding container 42292750073adc527222839210572c295bbb02e412cfb0d636ea4c27ac35a3c4: Status 404 returned error can't find the container with id 42292750073adc527222839210572c295bbb02e412cfb0d636ea4c27ac35a3c4 Feb 17 00:15:22 crc kubenswrapper[4690]: W0217 00:15:22.786110 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80f87711_62d4_488b_8a93_3dc6d93e3cc4.slice/crio-c49698f30be2c0265a2c9b7e4b458fa7da15dc78de11547a5ddfd00a4c12211f WatchSource:0}: Error finding container c49698f30be2c0265a2c9b7e4b458fa7da15dc78de11547a5ddfd00a4c12211f: Status 404 returned error can't find the container with id c49698f30be2c0265a2c9b7e4b458fa7da15dc78de11547a5ddfd00a4c12211f Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.792186 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" Feb 17 00:15:22 crc kubenswrapper[4690]: I0217 00:15:22.897529 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-7mmxh"] Feb 17 00:15:22 crc kubenswrapper[4690]: W0217 00:15:22.908072 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbae5780c_307b_4f82_8759_bde07af42405.slice/crio-145d8337c80b9d73f2b3374051d7e5d2337cde54eca3d9803cc7a684707f0a27 WatchSource:0}: Error finding container 145d8337c80b9d73f2b3374051d7e5d2337cde54eca3d9803cc7a684707f0a27: Status 404 returned error can't find the container with id 145d8337c80b9d73f2b3374051d7e5d2337cde54eca3d9803cc7a684707f0a27 Feb 17 00:15:23 crc kubenswrapper[4690]: I0217 00:15:23.023282 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-dj5w8"] Feb 17 00:15:23 crc kubenswrapper[4690]: W0217 00:15:23.029179 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6e1fe6c_28db_455c_b460_827567af661d.slice/crio-8ffc46265758c87bc72adcb46635c96379258f847605b77b55861e0a61936c8a WatchSource:0}: Error finding container 8ffc46265758c87bc72adcb46635c96379258f847605b77b55861e0a61936c8a: Status 404 returned error can't find the container with id 8ffc46265758c87bc72adcb46635c96379258f847605b77b55861e0a61936c8a Feb 17 00:15:23 crc kubenswrapper[4690]: I0217 00:15:23.044841 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-69c756cfd-9qjz2"] Feb 17 00:15:23 crc kubenswrapper[4690]: W0217 00:15:23.048306 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e25f5a3_f7d6_48a6_b2f5_c085fcb47db7.slice/crio-5b4c85ca4b0a926f8f68d1e72866a812a7893d95bf51d2c52dbf9a6933cea680 WatchSource:0}: Error finding container 5b4c85ca4b0a926f8f68d1e72866a812a7893d95bf51d2c52dbf9a6933cea680: Status 404 returned error can't find the container with id 5b4c85ca4b0a926f8f68d1e72866a812a7893d95bf51d2c52dbf9a6933cea680 Feb 17 00:15:23 crc kubenswrapper[4690]: I0217 00:15:23.734094 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" event={"ID":"d6e1fe6c-28db-455c-b460-827567af661d","Type":"ContainerStarted","Data":"8ffc46265758c87bc72adcb46635c96379258f847605b77b55861e0a61936c8a"} Feb 17 00:15:23 crc kubenswrapper[4690]: I0217 00:15:23.735141 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" event={"ID":"bae5780c-307b-4f82-8759-bde07af42405","Type":"ContainerStarted","Data":"145d8337c80b9d73f2b3374051d7e5d2337cde54eca3d9803cc7a684707f0a27"} Feb 17 00:15:23 crc kubenswrapper[4690]: I0217 00:15:23.736206 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" event={"ID":"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7","Type":"ContainerStarted","Data":"5b4c85ca4b0a926f8f68d1e72866a812a7893d95bf51d2c52dbf9a6933cea680"} Feb 17 00:15:23 crc kubenswrapper[4690]: I0217 00:15:23.736918 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" event={"ID":"dab3686a-8cc4-4105-8460-a534c5d7558d","Type":"ContainerStarted","Data":"42292750073adc527222839210572c295bbb02e412cfb0d636ea4c27ac35a3c4"} Feb 17 00:15:23 crc kubenswrapper[4690]: I0217 00:15:23.737667 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb" event={"ID":"80f87711-62d4-488b-8a93-3dc6d93e3cc4","Type":"ContainerStarted","Data":"c49698f30be2c0265a2c9b7e4b458fa7da15dc78de11547a5ddfd00a4c12211f"} Feb 17 00:15:36 crc kubenswrapper[4690]: E0217 00:15:36.458884 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8" Feb 17 00:15:36 crc kubenswrapper[4690]: E0217 00:15:36.459633 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xjf25,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5bf474d74f-dj5w8_openshift-operators(d6e1fe6c-28db-455c-b460-827567af661d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 17 00:15:36 crc kubenswrapper[4690]: E0217 00:15:36.461995 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" podUID="d6e1fe6c-28db-455c-b460-827567af661d" Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.835142 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" event={"ID":"bae5780c-307b-4f82-8759-bde07af42405","Type":"ContainerStarted","Data":"bc685394ad80ec68761f1440d3b47876d2ff2c8ae9974d3cd0868fbbab0e620b"} Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.835532 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.836866 4690 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-7mmxh container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.48:8081/healthz\": dial tcp 10.217.0.48:8081: connect: connection refused" start-of-body= Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.836930 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" podUID="bae5780c-307b-4f82-8759-bde07af42405" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.48:8081/healthz\": dial tcp 10.217.0.48:8081: connect: connection refused" Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.837277 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-86q9q" event={"ID":"4e01fa77-1ce1-4285-9845-722ccd914395","Type":"ContainerStarted","Data":"da6ddfb0a59a26101cf553603a9b096471b130eb25471c7f8c64f96f5b8e1055"} Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.838829 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" event={"ID":"5e25f5a3-f7d6-48a6-b2f5-c085fcb47db7","Type":"ContainerStarted","Data":"efb8a5bdfa3c90ea25f870c0bcff5eeca18a63a7e4d14c1f97b27f215fb12358"} Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.840531 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" event={"ID":"dab3686a-8cc4-4105-8460-a534c5d7558d","Type":"ContainerStarted","Data":"306cd5696a9500d20eb40e48b10193a2643bd714d4282d4e4416e5b7a0a59d1b"} Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.842995 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" event={"ID":"f4c6248a-697b-47eb-b7af-603a2bdf7ec0","Type":"ContainerStarted","Data":"76fe735fe3263e4d0f3a5291129fbc5b6cab65e8ca0ccf8bfc164d9ee0831d6f"} Feb 17 00:15:36 crc kubenswrapper[4690]: E0217 00:15:36.845836 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8\\\"\"" pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" podUID="d6e1fe6c-28db-455c-b460-827567af661d" Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.857786 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" podStartSLOduration=1.221200604 podStartE2EDuration="14.857766983s" podCreationTimestamp="2026-02-17 00:15:22 +0000 UTC" firstStartedPulling="2026-02-17 00:15:22.9108832 +0000 UTC m=+596.646595951" lastFinishedPulling="2026-02-17 00:15:36.547449579 +0000 UTC m=+610.283162330" observedRunningTime="2026-02-17 00:15:36.854514741 +0000 UTC m=+610.590227502" watchObservedRunningTime="2026-02-17 00:15:36.857766983 +0000 UTC m=+610.593479734" Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.892041 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-86q9q" podStartSLOduration=1.963885049 podStartE2EDuration="19.892021803s" podCreationTimestamp="2026-02-17 00:15:17 +0000 UTC" firstStartedPulling="2026-02-17 00:15:18.53432912 +0000 UTC m=+592.270041871" lastFinishedPulling="2026-02-17 00:15:36.462465884 +0000 UTC m=+610.198178625" observedRunningTime="2026-02-17 00:15:36.869883376 +0000 UTC m=+610.605596127" watchObservedRunningTime="2026-02-17 00:15:36.892021803 +0000 UTC m=+610.627734554" Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.894435 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb" podStartSLOduration=2.18892526 podStartE2EDuration="15.89441938s" podCreationTimestamp="2026-02-17 00:15:21 +0000 UTC" firstStartedPulling="2026-02-17 00:15:22.800952519 +0000 UTC m=+596.536665270" lastFinishedPulling="2026-02-17 00:15:36.506446639 +0000 UTC m=+610.242159390" observedRunningTime="2026-02-17 00:15:36.888209035 +0000 UTC m=+610.623921806" watchObservedRunningTime="2026-02-17 00:15:36.89441938 +0000 UTC m=+610.630132121" Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.906180 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj" podStartSLOduration=1.187353046 podStartE2EDuration="14.906162003s" podCreationTimestamp="2026-02-17 00:15:22 +0000 UTC" firstStartedPulling="2026-02-17 00:15:22.786135399 +0000 UTC m=+596.521848150" lastFinishedPulling="2026-02-17 00:15:36.504944356 +0000 UTC m=+610.240657107" observedRunningTime="2026-02-17 00:15:36.905797292 +0000 UTC m=+610.641510063" watchObservedRunningTime="2026-02-17 00:15:36.906162003 +0000 UTC m=+610.641874754" Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.961821 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-69c756cfd-9qjz2" podStartSLOduration=1.506610764 podStartE2EDuration="14.961798258s" podCreationTimestamp="2026-02-17 00:15:22 +0000 UTC" firstStartedPulling="2026-02-17 00:15:23.051607894 +0000 UTC m=+596.787320645" lastFinishedPulling="2026-02-17 00:15:36.506795388 +0000 UTC m=+610.242508139" observedRunningTime="2026-02-17 00:15:36.957889097 +0000 UTC m=+610.693601848" watchObservedRunningTime="2026-02-17 00:15:36.961798258 +0000 UTC m=+610.697511009" Feb 17 00:15:36 crc kubenswrapper[4690]: I0217 00:15:36.981396 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg" podStartSLOduration=1.138361829 podStartE2EDuration="14.981352101s" podCreationTimestamp="2026-02-17 00:15:22 +0000 UTC" firstStartedPulling="2026-02-17 00:15:22.663049845 +0000 UTC m=+596.398762596" lastFinishedPulling="2026-02-17 00:15:36.506040117 +0000 UTC m=+610.241752868" observedRunningTime="2026-02-17 00:15:36.976032731 +0000 UTC m=+610.711745492" watchObservedRunningTime="2026-02-17 00:15:36.981352101 +0000 UTC m=+610.717064852" Feb 17 00:15:37 crc kubenswrapper[4690]: I0217 00:15:37.852952 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-hjdqb" event={"ID":"80f87711-62d4-488b-8a93-3dc6d93e3cc4","Type":"ContainerStarted","Data":"8dda98bccc8de08c863534be0fc6658d52fa88ff46aef67c6d4674e2e515b49c"} Feb 17 00:15:37 crc kubenswrapper[4690]: I0217 00:15:37.854626 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-7mmxh" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.224303 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.225641 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.230461 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.230708 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.230925 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.230924 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.231560 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.232804 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.232899 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.233127 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-f9fnt" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.233499 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.246938 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.392923 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393042 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393484 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393533 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/c3743039-d680-4bec-9112-727ad3c494ad-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393615 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393641 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393667 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393687 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393707 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393727 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393743 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393764 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393849 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393893 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.393920 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494306 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494382 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494409 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494435 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/c3743039-d680-4bec-9112-727ad3c494ad-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494457 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494633 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494654 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494672 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494690 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494707 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494722 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494743 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494767 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494784 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.494806 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.495502 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.495966 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.496156 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.496240 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.496268 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.496279 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.496643 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.499583 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.506283 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.507485 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/c3743039-d680-4bec-9112-727ad3c494ad-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.507744 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.511998 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.512664 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.521246 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.521917 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/c3743039-d680-4bec-9112-727ad3c494ad-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"c3743039-d680-4bec-9112-727ad3c494ad\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.542214 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.806135 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 17 00:15:39 crc kubenswrapper[4690]: W0217 00:15:39.810104 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3743039_d680_4bec_9112_727ad3c494ad.slice/crio-fbf843d4b150fd98555dda2bd4fc54b638c7391c2a555185572685a5b9f580dc WatchSource:0}: Error finding container fbf843d4b150fd98555dda2bd4fc54b638c7391c2a555185572685a5b9f580dc: Status 404 returned error can't find the container with id fbf843d4b150fd98555dda2bd4fc54b638c7391c2a555185572685a5b9f580dc Feb 17 00:15:39 crc kubenswrapper[4690]: I0217 00:15:39.869807 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c3743039-d680-4bec-9112-727ad3c494ad","Type":"ContainerStarted","Data":"fbf843d4b150fd98555dda2bd4fc54b638c7391c2a555185572685a5b9f580dc"} Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.573305 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn"] Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.574269 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.576070 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.576438 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.576595 4690 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-4knvz" Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.586770 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn"] Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.719063 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwwc7\" (UniqueName: \"kubernetes.io/projected/afb39db1-a44e-47f6-ad2b-ae3a5f5a476c-kube-api-access-hwwc7\") pod \"cert-manager-operator-controller-manager-5586865c96-pdjfn\" (UID: \"afb39db1-a44e-47f6-ad2b-ae3a5f5a476c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.719146 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/afb39db1-a44e-47f6-ad2b-ae3a5f5a476c-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-pdjfn\" (UID: \"afb39db1-a44e-47f6-ad2b-ae3a5f5a476c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.820130 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwwc7\" (UniqueName: \"kubernetes.io/projected/afb39db1-a44e-47f6-ad2b-ae3a5f5a476c-kube-api-access-hwwc7\") pod \"cert-manager-operator-controller-manager-5586865c96-pdjfn\" (UID: \"afb39db1-a44e-47f6-ad2b-ae3a5f5a476c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.820210 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/afb39db1-a44e-47f6-ad2b-ae3a5f5a476c-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-pdjfn\" (UID: \"afb39db1-a44e-47f6-ad2b-ae3a5f5a476c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.820654 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/afb39db1-a44e-47f6-ad2b-ae3a5f5a476c-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-pdjfn\" (UID: \"afb39db1-a44e-47f6-ad2b-ae3a5f5a476c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.848047 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwwc7\" (UniqueName: \"kubernetes.io/projected/afb39db1-a44e-47f6-ad2b-ae3a5f5a476c-kube-api-access-hwwc7\") pod \"cert-manager-operator-controller-manager-5586865c96-pdjfn\" (UID: \"afb39db1-a44e-47f6-ad2b-ae3a5f5a476c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" Feb 17 00:15:41 crc kubenswrapper[4690]: I0217 00:15:41.890547 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" Feb 17 00:15:42 crc kubenswrapper[4690]: I0217 00:15:42.222807 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn"] Feb 17 00:15:42 crc kubenswrapper[4690]: I0217 00:15:42.898409 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" event={"ID":"afb39db1-a44e-47f6-ad2b-ae3a5f5a476c","Type":"ContainerStarted","Data":"19c4b2dcb5b3d57ed112e3e2040472e5297125147d1635abdc6e9a20564a09de"} Feb 17 00:15:55 crc kubenswrapper[4690]: I0217 00:15:55.983238 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" event={"ID":"d6e1fe6c-28db-455c-b460-827567af661d","Type":"ContainerStarted","Data":"9232ecf7a09704f56633dcf23821c80224947b2263ff87699f50facbc5e443fe"} Feb 17 00:15:55 crc kubenswrapper[4690]: I0217 00:15:55.984687 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" Feb 17 00:15:55 crc kubenswrapper[4690]: I0217 00:15:55.986647 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" event={"ID":"afb39db1-a44e-47f6-ad2b-ae3a5f5a476c","Type":"ContainerStarted","Data":"29ab4e347946565f3b5405ee6b9225347da5934fbb22872ea5adef73321c55f1"} Feb 17 00:15:55 crc kubenswrapper[4690]: I0217 00:15:55.988769 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c3743039-d680-4bec-9112-727ad3c494ad","Type":"ContainerStarted","Data":"97b8fe8ea8b07171e37c11e077a77864700e548152c0f4694c820864ea780ff8"} Feb 17 00:15:56 crc kubenswrapper[4690]: I0217 00:15:56.020223 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" podStartSLOduration=2.105588238 podStartE2EDuration="34.020190085s" podCreationTimestamp="2026-02-17 00:15:22 +0000 UTC" firstStartedPulling="2026-02-17 00:15:23.043019851 +0000 UTC m=+596.778732602" lastFinishedPulling="2026-02-17 00:15:54.957621658 +0000 UTC m=+628.693334449" observedRunningTime="2026-02-17 00:15:56.009239655 +0000 UTC m=+629.744952406" watchObservedRunningTime="2026-02-17 00:15:56.020190085 +0000 UTC m=+629.755902876" Feb 17 00:15:56 crc kubenswrapper[4690]: I0217 00:15:56.113806 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-pdjfn" podStartSLOduration=2.564093459 podStartE2EDuration="15.113781384s" podCreationTimestamp="2026-02-17 00:15:41 +0000 UTC" firstStartedPulling="2026-02-17 00:15:42.247560164 +0000 UTC m=+615.983272905" lastFinishedPulling="2026-02-17 00:15:54.797248049 +0000 UTC m=+628.532960830" observedRunningTime="2026-02-17 00:15:56.111451948 +0000 UTC m=+629.847164729" watchObservedRunningTime="2026-02-17 00:15:56.113781384 +0000 UTC m=+629.849494175" Feb 17 00:15:56 crc kubenswrapper[4690]: I0217 00:15:56.189898 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 17 00:15:56 crc kubenswrapper[4690]: I0217 00:15:56.227649 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 17 00:15:58 crc kubenswrapper[4690]: I0217 00:15:58.000342 4690 generic.go:334] "Generic (PLEG): container finished" podID="c3743039-d680-4bec-9112-727ad3c494ad" containerID="97b8fe8ea8b07171e37c11e077a77864700e548152c0f4694c820864ea780ff8" exitCode=0 Feb 17 00:15:58 crc kubenswrapper[4690]: I0217 00:15:58.000384 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c3743039-d680-4bec-9112-727ad3c494ad","Type":"ContainerDied","Data":"97b8fe8ea8b07171e37c11e077a77864700e548152c0f4694c820864ea780ff8"} Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.009936 4690 generic.go:334] "Generic (PLEG): container finished" podID="c3743039-d680-4bec-9112-727ad3c494ad" containerID="0abe12a34ef68e038e92a202c76063ee53e13f9caf284ad240dfc2639067a68f" exitCode=0 Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.010024 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c3743039-d680-4bec-9112-727ad3c494ad","Type":"ContainerDied","Data":"0abe12a34ef68e038e92a202c76063ee53e13f9caf284ad240dfc2639067a68f"} Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.682063 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-58lzf"] Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.686567 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.692604 4690 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-c7p5b" Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.692639 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.694818 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-58lzf"] Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.710123 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.822554 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9ca0053-6e23-41ba-b825-80acdae13e89-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-58lzf\" (UID: \"e9ca0053-6e23-41ba-b825-80acdae13e89\") " pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.822638 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84dff\" (UniqueName: \"kubernetes.io/projected/e9ca0053-6e23-41ba-b825-80acdae13e89-kube-api-access-84dff\") pod \"cert-manager-webhook-6888856db4-58lzf\" (UID: \"e9ca0053-6e23-41ba-b825-80acdae13e89\") " pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.924010 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9ca0053-6e23-41ba-b825-80acdae13e89-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-58lzf\" (UID: \"e9ca0053-6e23-41ba-b825-80acdae13e89\") " pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.924121 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84dff\" (UniqueName: \"kubernetes.io/projected/e9ca0053-6e23-41ba-b825-80acdae13e89-kube-api-access-84dff\") pod \"cert-manager-webhook-6888856db4-58lzf\" (UID: \"e9ca0053-6e23-41ba-b825-80acdae13e89\") " pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.944211 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9ca0053-6e23-41ba-b825-80acdae13e89-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-58lzf\" (UID: \"e9ca0053-6e23-41ba-b825-80acdae13e89\") " pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" Feb 17 00:15:59 crc kubenswrapper[4690]: I0217 00:15:59.947476 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84dff\" (UniqueName: \"kubernetes.io/projected/e9ca0053-6e23-41ba-b825-80acdae13e89-kube-api-access-84dff\") pod \"cert-manager-webhook-6888856db4-58lzf\" (UID: \"e9ca0053-6e23-41ba-b825-80acdae13e89\") " pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.017920 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c3743039-d680-4bec-9112-727ad3c494ad","Type":"ContainerStarted","Data":"4202bc5810b2a59e35f568e5ae92d2ea4f80958aeb2345281be6f7ee77e7d242"} Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.018579 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.044845 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=5.756715356 podStartE2EDuration="21.044825194s" podCreationTimestamp="2026-02-17 00:15:39 +0000 UTC" firstStartedPulling="2026-02-17 00:15:39.811952612 +0000 UTC m=+613.547665363" lastFinishedPulling="2026-02-17 00:15:55.10006245 +0000 UTC m=+628.835775201" observedRunningTime="2026-02-17 00:16:00.042653512 +0000 UTC m=+633.778366273" watchObservedRunningTime="2026-02-17 00:16:00.044825194 +0000 UTC m=+633.780537945" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.047989 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.539617 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-58lzf"] Feb 17 00:16:00 crc kubenswrapper[4690]: W0217 00:16:00.544381 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9ca0053_6e23_41ba_b825_80acdae13e89.slice/crio-465b00d9a176933e8d523a9fa71fc292e79b3df6f4c367c087ffab969032430e WatchSource:0}: Error finding container 465b00d9a176933e8d523a9fa71fc292e79b3df6f4c367c087ffab969032430e: Status 404 returned error can't find the container with id 465b00d9a176933e8d523a9fa71fc292e79b3df6f4c367c087ffab969032430e Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.776780 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-wtxfw"] Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.778167 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.780291 4690 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-dmqd4" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.784394 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-wtxfw"] Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.834492 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7728f7bd-93da-484e-b9df-2f4834cbd19b-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-wtxfw\" (UID: \"7728f7bd-93da-484e-b9df-2f4834cbd19b\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.834549 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbxd5\" (UniqueName: \"kubernetes.io/projected/7728f7bd-93da-484e-b9df-2f4834cbd19b-kube-api-access-qbxd5\") pod \"cert-manager-cainjector-5545bd876-wtxfw\" (UID: \"7728f7bd-93da-484e-b9df-2f4834cbd19b\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.936594 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7728f7bd-93da-484e-b9df-2f4834cbd19b-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-wtxfw\" (UID: \"7728f7bd-93da-484e-b9df-2f4834cbd19b\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.936631 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbxd5\" (UniqueName: \"kubernetes.io/projected/7728f7bd-93da-484e-b9df-2f4834cbd19b-kube-api-access-qbxd5\") pod \"cert-manager-cainjector-5545bd876-wtxfw\" (UID: \"7728f7bd-93da-484e-b9df-2f4834cbd19b\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.955309 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7728f7bd-93da-484e-b9df-2f4834cbd19b-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-wtxfw\" (UID: \"7728f7bd-93da-484e-b9df-2f4834cbd19b\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" Feb 17 00:16:00 crc kubenswrapper[4690]: I0217 00:16:00.955691 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbxd5\" (UniqueName: \"kubernetes.io/projected/7728f7bd-93da-484e-b9df-2f4834cbd19b-kube-api-access-qbxd5\") pod \"cert-manager-cainjector-5545bd876-wtxfw\" (UID: \"7728f7bd-93da-484e-b9df-2f4834cbd19b\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" Feb 17 00:16:01 crc kubenswrapper[4690]: I0217 00:16:01.030291 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" event={"ID":"e9ca0053-6e23-41ba-b825-80acdae13e89","Type":"ContainerStarted","Data":"465b00d9a176933e8d523a9fa71fc292e79b3df6f4c367c087ffab969032430e"} Feb 17 00:16:01 crc kubenswrapper[4690]: I0217 00:16:01.100417 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" Feb 17 00:16:01 crc kubenswrapper[4690]: I0217 00:16:01.335942 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-wtxfw"] Feb 17 00:16:01 crc kubenswrapper[4690]: W0217 00:16:01.337705 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7728f7bd_93da_484e_b9df_2f4834cbd19b.slice/crio-a88063978f4a3ebbd6522b92924251a6ed4c230fb878350ad980aa9551d1adab WatchSource:0}: Error finding container a88063978f4a3ebbd6522b92924251a6ed4c230fb878350ad980aa9551d1adab: Status 404 returned error can't find the container with id a88063978f4a3ebbd6522b92924251a6ed4c230fb878350ad980aa9551d1adab Feb 17 00:16:02 crc kubenswrapper[4690]: I0217 00:16:02.036018 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" event={"ID":"7728f7bd-93da-484e-b9df-2f4834cbd19b","Type":"ContainerStarted","Data":"a88063978f4a3ebbd6522b92924251a6ed4c230fb878350ad980aa9551d1adab"} Feb 17 00:16:02 crc kubenswrapper[4690]: I0217 00:16:02.796065 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-dj5w8" Feb 17 00:16:08 crc kubenswrapper[4690]: I0217 00:16:08.076345 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" event={"ID":"e9ca0053-6e23-41ba-b825-80acdae13e89","Type":"ContainerStarted","Data":"cca4515f59f8789e91e2716d03dd0d8d05692cd4cc9bc8a576165c88d01924b7"} Feb 17 00:16:08 crc kubenswrapper[4690]: I0217 00:16:08.076792 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" Feb 17 00:16:08 crc kubenswrapper[4690]: I0217 00:16:08.078022 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" event={"ID":"7728f7bd-93da-484e-b9df-2f4834cbd19b","Type":"ContainerStarted","Data":"aa9b901bda689aa7ddb95d06bd5ef0ee6b52b35966c9b770ab3b5125c14be9d2"} Feb 17 00:16:08 crc kubenswrapper[4690]: I0217 00:16:08.107352 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" podStartSLOduration=2.656115922 podStartE2EDuration="9.107338477s" podCreationTimestamp="2026-02-17 00:15:59 +0000 UTC" firstStartedPulling="2026-02-17 00:16:00.548554202 +0000 UTC m=+634.284266953" lastFinishedPulling="2026-02-17 00:16:06.999776747 +0000 UTC m=+640.735489508" observedRunningTime="2026-02-17 00:16:08.105488555 +0000 UTC m=+641.841201306" watchObservedRunningTime="2026-02-17 00:16:08.107338477 +0000 UTC m=+641.843051228" Feb 17 00:16:08 crc kubenswrapper[4690]: I0217 00:16:08.125050 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-wtxfw" podStartSLOduration=2.450140587 podStartE2EDuration="8.125038448s" podCreationTimestamp="2026-02-17 00:16:00 +0000 UTC" firstStartedPulling="2026-02-17 00:16:01.343315327 +0000 UTC m=+635.079028078" lastFinishedPulling="2026-02-17 00:16:07.018213188 +0000 UTC m=+640.753925939" observedRunningTime="2026-02-17 00:16:08.124458922 +0000 UTC m=+641.860171673" watchObservedRunningTime="2026-02-17 00:16:08.125038448 +0000 UTC m=+641.860751199" Feb 17 00:16:09 crc kubenswrapper[4690]: I0217 00:16:09.646139 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="c3743039-d680-4bec-9112-727ad3c494ad" containerName="elasticsearch" probeResult="failure" output=< Feb 17 00:16:09 crc kubenswrapper[4690]: {"timestamp": "2026-02-17T00:16:09+00:00", "message": "readiness probe failed", "curl_rc": "7"} Feb 17 00:16:09 crc kubenswrapper[4690]: > Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.575180 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.576053 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.577625 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-nkjdw" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.578475 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.578492 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.578631 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.598988 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.672988 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673038 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673083 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673106 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673129 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673145 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673163 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673200 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673218 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673234 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcvvv\" (UniqueName: \"kubernetes.io/projected/e98a4253-41c0-49d4-b47f-a60446334be6-kube-api-access-fcvvv\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673253 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.673276 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774113 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774157 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774188 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774208 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774232 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774249 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774269 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774302 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774321 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774342 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcvvv\" (UniqueName: \"kubernetes.io/projected/e98a4253-41c0-49d4-b47f-a60446334be6-kube-api-access-fcvvv\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774399 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774421 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774773 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774397 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774815 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.774866 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.775150 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.775297 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.775306 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.775391 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.775429 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.782426 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.784844 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.796215 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcvvv\" (UniqueName: \"kubernetes.io/projected/e98a4253-41c0-49d4-b47f-a60446334be6-kube-api-access-fcvvv\") pod \"service-telemetry-operator-1-build\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:11 crc kubenswrapper[4690]: I0217 00:16:11.889293 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:12 crc kubenswrapper[4690]: I0217 00:16:12.135773 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 17 00:16:13 crc kubenswrapper[4690]: I0217 00:16:13.106325 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"e98a4253-41c0-49d4-b47f-a60446334be6","Type":"ContainerStarted","Data":"85ba6735636b4d9c879a11ffbf03a400ea1d29f29623e79ac043a679813c5d80"} Feb 17 00:16:15 crc kubenswrapper[4690]: I0217 00:16:15.031286 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Feb 17 00:16:15 crc kubenswrapper[4690]: I0217 00:16:15.050730 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-58lzf" Feb 17 00:16:17 crc kubenswrapper[4690]: I0217 00:16:17.963554 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-kpkt9"] Feb 17 00:16:17 crc kubenswrapper[4690]: I0217 00:16:17.964702 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-kpkt9" Feb 17 00:16:17 crc kubenswrapper[4690]: I0217 00:16:17.974723 4690 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-b65tl" Feb 17 00:16:17 crc kubenswrapper[4690]: I0217 00:16:17.979921 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-kpkt9"] Feb 17 00:16:18 crc kubenswrapper[4690]: I0217 00:16:18.131223 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crctj\" (UniqueName: \"kubernetes.io/projected/01b55f46-e76a-4a56-a49c-888209f5b626-kube-api-access-crctj\") pod \"cert-manager-545d4d4674-kpkt9\" (UID: \"01b55f46-e76a-4a56-a49c-888209f5b626\") " pod="cert-manager/cert-manager-545d4d4674-kpkt9" Feb 17 00:16:18 crc kubenswrapper[4690]: I0217 00:16:18.131334 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01b55f46-e76a-4a56-a49c-888209f5b626-bound-sa-token\") pod \"cert-manager-545d4d4674-kpkt9\" (UID: \"01b55f46-e76a-4a56-a49c-888209f5b626\") " pod="cert-manager/cert-manager-545d4d4674-kpkt9" Feb 17 00:16:18 crc kubenswrapper[4690]: I0217 00:16:18.232862 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crctj\" (UniqueName: \"kubernetes.io/projected/01b55f46-e76a-4a56-a49c-888209f5b626-kube-api-access-crctj\") pod \"cert-manager-545d4d4674-kpkt9\" (UID: \"01b55f46-e76a-4a56-a49c-888209f5b626\") " pod="cert-manager/cert-manager-545d4d4674-kpkt9" Feb 17 00:16:18 crc kubenswrapper[4690]: I0217 00:16:18.232951 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01b55f46-e76a-4a56-a49c-888209f5b626-bound-sa-token\") pod \"cert-manager-545d4d4674-kpkt9\" (UID: \"01b55f46-e76a-4a56-a49c-888209f5b626\") " pod="cert-manager/cert-manager-545d4d4674-kpkt9" Feb 17 00:16:18 crc kubenswrapper[4690]: I0217 00:16:18.257619 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01b55f46-e76a-4a56-a49c-888209f5b626-bound-sa-token\") pod \"cert-manager-545d4d4674-kpkt9\" (UID: \"01b55f46-e76a-4a56-a49c-888209f5b626\") " pod="cert-manager/cert-manager-545d4d4674-kpkt9" Feb 17 00:16:18 crc kubenswrapper[4690]: I0217 00:16:18.262239 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crctj\" (UniqueName: \"kubernetes.io/projected/01b55f46-e76a-4a56-a49c-888209f5b626-kube-api-access-crctj\") pod \"cert-manager-545d4d4674-kpkt9\" (UID: \"01b55f46-e76a-4a56-a49c-888209f5b626\") " pod="cert-manager/cert-manager-545d4d4674-kpkt9" Feb 17 00:16:18 crc kubenswrapper[4690]: I0217 00:16:18.300944 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-kpkt9" Feb 17 00:16:18 crc kubenswrapper[4690]: I0217 00:16:18.775975 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-kpkt9"] Feb 17 00:16:18 crc kubenswrapper[4690]: I0217 00:16:18.853258 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-kpkt9" event={"ID":"01b55f46-e76a-4a56-a49c-888209f5b626","Type":"ContainerStarted","Data":"88bb27a405619e62f7a251842ae43a50932a51b1f59e146fe734b7667173c28d"} Feb 17 00:16:19 crc kubenswrapper[4690]: I0217 00:16:19.872321 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-kpkt9" event={"ID":"01b55f46-e76a-4a56-a49c-888209f5b626","Type":"ContainerStarted","Data":"fc06af7ebc174dcf854376e6e01e439936b2467a5354b85519c72499bdf44a18"} Feb 17 00:16:19 crc kubenswrapper[4690]: I0217 00:16:19.887930 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-kpkt9" podStartSLOduration=2.887909421 podStartE2EDuration="2.887909421s" podCreationTimestamp="2026-02-17 00:16:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 00:16:19.887226212 +0000 UTC m=+653.622938963" watchObservedRunningTime="2026-02-17 00:16:19.887909421 +0000 UTC m=+653.623622212" Feb 17 00:16:22 crc kubenswrapper[4690]: I0217 00:16:22.077285 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.679417 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.681810 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.683656 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.686827 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.687087 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.704769 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.712164 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.712291 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.712354 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m92x\" (UniqueName: \"kubernetes.io/projected/8075fea6-4804-43bf-b604-c401bde1c9c3-kube-api-access-7m92x\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.712709 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.712870 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.712985 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.713035 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.713114 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.713201 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.713309 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.713438 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.713611 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.815627 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.815702 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.815748 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.815784 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m92x\" (UniqueName: \"kubernetes.io/projected/8075fea6-4804-43bf-b604-c401bde1c9c3-kube-api-access-7m92x\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.815875 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.815924 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.815966 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.816006 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.816040 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.816074 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.816112 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.816149 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.816291 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.816408 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.816456 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.816864 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.816918 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.817397 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.818130 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.818248 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.819095 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.824591 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.825824 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.846714 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m92x\" (UniqueName: \"kubernetes.io/projected/8075fea6-4804-43bf-b604-c401bde1c9c3-kube-api-access-7m92x\") pod \"service-telemetry-operator-2-build\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.900421 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"e98a4253-41c0-49d4-b47f-a60446334be6","Type":"ContainerStarted","Data":"18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50"} Feb 17 00:16:23 crc kubenswrapper[4690]: I0217 00:16:23.900705 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="e98a4253-41c0-49d4-b47f-a60446334be6" containerName="manage-dockerfile" containerID="cri-o://18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50" gracePeriod=30 Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.006601 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.271439 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.287225 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_e98a4253-41c0-49d4-b47f-a60446334be6/manage-dockerfile/0.log" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.287288 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:24 crc kubenswrapper[4690]: W0217 00:16:24.287950 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8075fea6_4804_43bf_b604_c401bde1c9c3.slice/crio-a23d687f5c5afacb6bd3ad87a20bdca295952973b94566775587240b11263aa1 WatchSource:0}: Error finding container a23d687f5c5afacb6bd3ad87a20bdca295952973b94566775587240b11263aa1: Status 404 returned error can't find the container with id a23d687f5c5afacb6bd3ad87a20bdca295952973b94566775587240b11263aa1 Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.424802 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425116 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-root\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425158 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-run\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425195 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-proxy-ca-bundles\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425219 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-system-configs\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425246 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-buildworkdir\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425272 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcvvv\" (UniqueName: \"kubernetes.io/projected/e98a4253-41c0-49d4-b47f-a60446334be6-kube-api-access-fcvvv\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425296 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-ca-bundles\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425326 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-node-pullsecrets\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425348 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-build-blob-cache\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425385 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-pull\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425420 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-push\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425444 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-buildcachedir\") pod \"e98a4253-41c0-49d4-b47f-a60446334be6\" (UID: \"e98a4253-41c0-49d4-b47f-a60446334be6\") " Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425894 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.425936 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.426110 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.426808 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427111 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427512 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427550 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427783 4690 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427805 4690 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427817 4690 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427828 4690 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427842 4690 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427852 4690 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427854 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.427864 4690 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e98a4253-41c0-49d4-b47f-a60446334be6-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.430245 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-pull" (OuterVolumeSpecName: "builder-dockercfg-nkjdw-pull") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "builder-dockercfg-nkjdw-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.430314 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-push" (OuterVolumeSpecName: "builder-dockercfg-nkjdw-push") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "builder-dockercfg-nkjdw-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.430316 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e98a4253-41c0-49d4-b47f-a60446334be6-kube-api-access-fcvvv" (OuterVolumeSpecName: "kube-api-access-fcvvv") pod "e98a4253-41c0-49d4-b47f-a60446334be6" (UID: "e98a4253-41c0-49d4-b47f-a60446334be6"). InnerVolumeSpecName "kube-api-access-fcvvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.529026 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcvvv\" (UniqueName: \"kubernetes.io/projected/e98a4253-41c0-49d4-b47f-a60446334be6-kube-api-access-fcvvv\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.529052 4690 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-pull\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.529062 4690 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/e98a4253-41c0-49d4-b47f-a60446334be6-builder-dockercfg-nkjdw-push\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.529070 4690 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e98a4253-41c0-49d4-b47f-a60446334be6-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.529080 4690 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e98a4253-41c0-49d4-b47f-a60446334be6-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.906558 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"8075fea6-4804-43bf-b604-c401bde1c9c3","Type":"ContainerStarted","Data":"42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051"} Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.906600 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"8075fea6-4804-43bf-b604-c401bde1c9c3","Type":"ContainerStarted","Data":"a23d687f5c5afacb6bd3ad87a20bdca295952973b94566775587240b11263aa1"} Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.908449 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_e98a4253-41c0-49d4-b47f-a60446334be6/manage-dockerfile/0.log" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.908493 4690 generic.go:334] "Generic (PLEG): container finished" podID="e98a4253-41c0-49d4-b47f-a60446334be6" containerID="18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50" exitCode=1 Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.908520 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"e98a4253-41c0-49d4-b47f-a60446334be6","Type":"ContainerDied","Data":"18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50"} Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.908543 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"e98a4253-41c0-49d4-b47f-a60446334be6","Type":"ContainerDied","Data":"85ba6735636b4d9c879a11ffbf03a400ea1d29f29623e79ac043a679813c5d80"} Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.908558 4690 scope.go:117] "RemoveContainer" containerID="18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.908578 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.938564 4690 scope.go:117] "RemoveContainer" containerID="18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50" Feb 17 00:16:24 crc kubenswrapper[4690]: E0217 00:16:24.940940 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50\": container with ID starting with 18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50 not found: ID does not exist" containerID="18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.940987 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50"} err="failed to get container status \"18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50\": rpc error: code = NotFound desc = could not find container \"18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50\": container with ID starting with 18c6a41f083212571533ad13ef921c2566bd8e4e6c6a66d1881503ed2da00e50 not found: ID does not exist" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.960009 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 17 00:16:24 crc kubenswrapper[4690]: E0217 00:16:24.964839 4690 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=6832776695153960605, SKID=, AKID=F6:B0:84:3E:CE:5D:F7:B6:01:CF:ED:4A:AE:12:EA:43:DB:35:C5:B4 failed: x509: certificate signed by unknown authority" Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.964999 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 17 00:16:24 crc kubenswrapper[4690]: I0217 00:16:24.983744 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e98a4253-41c0-49d4-b47f-a60446334be6" path="/var/lib/kubelet/pods/e98a4253-41c0-49d4-b47f-a60446334be6/volumes" Feb 17 00:16:26 crc kubenswrapper[4690]: I0217 00:16:26.001106 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 17 00:16:26 crc kubenswrapper[4690]: I0217 00:16:26.925077 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-2-build" podUID="8075fea6-4804-43bf-b604-c401bde1c9c3" containerName="git-clone" containerID="cri-o://42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051" gracePeriod=30 Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.437138 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_8075fea6-4804-43bf-b604-c401bde1c9c3/git-clone/0.log" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.438471 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.478995 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-proxy-ca-bundles\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.479846 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-build-blob-cache\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.480268 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-push\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.480935 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-node-pullsecrets\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.481068 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-root\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.481174 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m92x\" (UniqueName: \"kubernetes.io/projected/8075fea6-4804-43bf-b604-c401bde1c9c3-kube-api-access-7m92x\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.481303 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-ca-bundles\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.481512 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-buildcachedir\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.482131 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-run\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.482284 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-system-configs\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.482420 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-buildworkdir\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.482521 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-pull\") pod \"8075fea6-4804-43bf-b604-c401bde1c9c3\" (UID: \"8075fea6-4804-43bf-b604-c401bde1c9c3\") " Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.479745 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.480210 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.481026 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.481417 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.481576 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.481845 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.482312 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.482528 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.482638 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.485494 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-pull" (OuterVolumeSpecName: "builder-dockercfg-nkjdw-pull") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "builder-dockercfg-nkjdw-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.485573 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-push" (OuterVolumeSpecName: "builder-dockercfg-nkjdw-push") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "builder-dockercfg-nkjdw-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.487519 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8075fea6-4804-43bf-b604-c401bde1c9c3-kube-api-access-7m92x" (OuterVolumeSpecName: "kube-api-access-7m92x") pod "8075fea6-4804-43bf-b604-c401bde1c9c3" (UID: "8075fea6-4804-43bf-b604-c401bde1c9c3"). InnerVolumeSpecName "kube-api-access-7m92x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584015 4690 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584054 4690 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584066 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m92x\" (UniqueName: \"kubernetes.io/projected/8075fea6-4804-43bf-b604-c401bde1c9c3-kube-api-access-7m92x\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584075 4690 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584083 4690 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8075fea6-4804-43bf-b604-c401bde1c9c3-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584090 4690 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584099 4690 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584107 4690 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584117 4690 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-pull\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584128 4690 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8075fea6-4804-43bf-b604-c401bde1c9c3-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584136 4690 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8075fea6-4804-43bf-b604-c401bde1c9c3-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.584146 4690 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/8075fea6-4804-43bf-b604-c401bde1c9c3-builder-dockercfg-nkjdw-push\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.936980 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_8075fea6-4804-43bf-b604-c401bde1c9c3/git-clone/0.log" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.937262 4690 generic.go:334] "Generic (PLEG): container finished" podID="8075fea6-4804-43bf-b604-c401bde1c9c3" containerID="42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051" exitCode=1 Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.937408 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.937439 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"8075fea6-4804-43bf-b604-c401bde1c9c3","Type":"ContainerDied","Data":"42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051"} Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.938086 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"8075fea6-4804-43bf-b604-c401bde1c9c3","Type":"ContainerDied","Data":"a23d687f5c5afacb6bd3ad87a20bdca295952973b94566775587240b11263aa1"} Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.938155 4690 scope.go:117] "RemoveContainer" containerID="42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.983805 4690 scope.go:117] "RemoveContainer" containerID="42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051" Feb 17 00:16:27 crc kubenswrapper[4690]: E0217 00:16:27.984841 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051\": container with ID starting with 42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051 not found: ID does not exist" containerID="42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.984901 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051"} err="failed to get container status \"42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051\": rpc error: code = NotFound desc = could not find container \"42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051\": container with ID starting with 42f6789313313a6c50c1e111ff7ff20ee1b7ee64a0262a2f6008164e61607051 not found: ID does not exist" Feb 17 00:16:27 crc kubenswrapper[4690]: I0217 00:16:27.993691 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 17 00:16:28 crc kubenswrapper[4690]: I0217 00:16:28.005132 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 17 00:16:28 crc kubenswrapper[4690]: I0217 00:16:28.992508 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8075fea6-4804-43bf-b604-c401bde1c9c3" path="/var/lib/kubelet/pods/8075fea6-4804-43bf-b604-c401bde1c9c3/volumes" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.435737 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 17 00:16:37 crc kubenswrapper[4690]: E0217 00:16:37.436689 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8075fea6-4804-43bf-b604-c401bde1c9c3" containerName="git-clone" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.436710 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8075fea6-4804-43bf-b604-c401bde1c9c3" containerName="git-clone" Feb 17 00:16:37 crc kubenswrapper[4690]: E0217 00:16:37.436730 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98a4253-41c0-49d4-b47f-a60446334be6" containerName="manage-dockerfile" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.436745 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98a4253-41c0-49d4-b47f-a60446334be6" containerName="manage-dockerfile" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.436948 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="8075fea6-4804-43bf-b604-c401bde1c9c3" containerName="git-clone" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.436981 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98a4253-41c0-49d4-b47f-a60446334be6" containerName="manage-dockerfile" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.438422 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.441028 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-global-ca" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.441176 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-nkjdw" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.441275 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-sys-config" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.445822 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-ca" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.511959 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518172 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518210 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518248 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv5mw\" (UniqueName: \"kubernetes.io/projected/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-kube-api-access-bv5mw\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518270 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518475 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518617 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518666 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518704 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518755 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518790 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518821 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.518881 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619486 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619577 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619607 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619634 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619655 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619681 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619703 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619736 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619767 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619790 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619821 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv5mw\" (UniqueName: \"kubernetes.io/projected/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-kube-api-access-bv5mw\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619842 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.619888 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.620029 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.620267 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.620467 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.620492 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.620543 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.620569 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.620774 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.621127 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.625477 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.626030 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.639281 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv5mw\" (UniqueName: \"kubernetes.io/projected/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-kube-api-access-bv5mw\") pod \"service-telemetry-operator-3-build\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:37 crc kubenswrapper[4690]: I0217 00:16:37.771040 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:38 crc kubenswrapper[4690]: I0217 00:16:38.033908 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 17 00:16:39 crc kubenswrapper[4690]: I0217 00:16:39.021166 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"7fa318b3-0f9c-475a-bdcf-3ade93525dcb","Type":"ContainerStarted","Data":"bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6"} Feb 17 00:16:39 crc kubenswrapper[4690]: I0217 00:16:39.021504 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"7fa318b3-0f9c-475a-bdcf-3ade93525dcb","Type":"ContainerStarted","Data":"6294ed81e45c624f9aace5d217791cd6c268199db7d7b5fbe9754ae44769ef50"} Feb 17 00:16:39 crc kubenswrapper[4690]: E0217 00:16:39.108679 4690 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=6832776695153960605, SKID=, AKID=F6:B0:84:3E:CE:5D:F7:B6:01:CF:ED:4A:AE:12:EA:43:DB:35:C5:B4 failed: x509: certificate signed by unknown authority" Feb 17 00:16:40 crc kubenswrapper[4690]: I0217 00:16:40.142316 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.035380 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-3-build" podUID="7fa318b3-0f9c-475a-bdcf-3ade93525dcb" containerName="git-clone" containerID="cri-o://bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6" gracePeriod=30 Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.413698 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-3-build_7fa318b3-0f9c-475a-bdcf-3ade93525dcb/git-clone/0.log" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.413768 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.502746 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv5mw\" (UniqueName: \"kubernetes.io/projected/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-kube-api-access-bv5mw\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.502799 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-push\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.502841 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-pull\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.502877 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-ca-bundles\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.502899 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-system-configs\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.502921 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-root\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.502936 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-proxy-ca-bundles\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.502994 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-run\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503015 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildworkdir\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503060 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildcachedir\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503079 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-blob-cache\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503102 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-node-pullsecrets\") pod \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\" (UID: \"7fa318b3-0f9c-475a-bdcf-3ade93525dcb\") " Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503152 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503216 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503461 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503471 4690 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503517 4690 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503525 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503676 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503819 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.503906 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.506687 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.509083 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-push" (OuterVolumeSpecName: "builder-dockercfg-nkjdw-push") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "builder-dockercfg-nkjdw-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.510066 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.512494 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-kube-api-access-bv5mw" (OuterVolumeSpecName: "kube-api-access-bv5mw") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "kube-api-access-bv5mw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.516477 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-pull" (OuterVolumeSpecName: "builder-dockercfg-nkjdw-pull") pod "7fa318b3-0f9c-475a-bdcf-3ade93525dcb" (UID: "7fa318b3-0f9c-475a-bdcf-3ade93525dcb"). InnerVolumeSpecName "builder-dockercfg-nkjdw-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.604753 4690 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-push\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.604788 4690 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-builder-dockercfg-nkjdw-pull\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.604799 4690 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.604808 4690 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.604818 4690 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.604826 4690 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.604836 4690 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.604844 4690 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.604851 4690 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:41 crc kubenswrapper[4690]: I0217 00:16:41.604859 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv5mw\" (UniqueName: \"kubernetes.io/projected/7fa318b3-0f9c-475a-bdcf-3ade93525dcb-kube-api-access-bv5mw\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.042450 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-3-build_7fa318b3-0f9c-475a-bdcf-3ade93525dcb/git-clone/0.log" Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.042514 4690 generic.go:334] "Generic (PLEG): container finished" podID="7fa318b3-0f9c-475a-bdcf-3ade93525dcb" containerID="bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6" exitCode=1 Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.042556 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"7fa318b3-0f9c-475a-bdcf-3ade93525dcb","Type":"ContainerDied","Data":"bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6"} Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.042591 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"7fa318b3-0f9c-475a-bdcf-3ade93525dcb","Type":"ContainerDied","Data":"6294ed81e45c624f9aace5d217791cd6c268199db7d7b5fbe9754ae44769ef50"} Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.042620 4690 scope.go:117] "RemoveContainer" containerID="bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6" Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.042824 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.061636 4690 scope.go:117] "RemoveContainer" containerID="bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6" Feb 17 00:16:42 crc kubenswrapper[4690]: E0217 00:16:42.062101 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6\": container with ID starting with bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6 not found: ID does not exist" containerID="bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6" Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.062146 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6"} err="failed to get container status \"bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6\": rpc error: code = NotFound desc = could not find container \"bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6\": container with ID starting with bfd306420f05476e7c2561f7fb98ae28a0ad3bfbd55e411da3376486635ba0a6 not found: ID does not exist" Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.087463 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.095271 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 17 00:16:42 crc kubenswrapper[4690]: I0217 00:16:42.986216 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fa318b3-0f9c-475a-bdcf-3ade93525dcb" path="/var/lib/kubelet/pods/7fa318b3-0f9c-475a-bdcf-3ade93525dcb/volumes" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.537450 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 17 00:16:51 crc kubenswrapper[4690]: E0217 00:16:51.538291 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa318b3-0f9c-475a-bdcf-3ade93525dcb" containerName="git-clone" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.538310 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa318b3-0f9c-475a-bdcf-3ade93525dcb" containerName="git-clone" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.538515 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fa318b3-0f9c-475a-bdcf-3ade93525dcb" containerName="git-clone" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.539751 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.542837 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-global-ca" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.543299 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-nkjdw" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.544490 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-sys-config" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546158 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546224 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546273 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546308 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546408 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546470 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546535 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g47wb\" (UniqueName: \"kubernetes.io/projected/9614d2ac-beb6-461b-99cf-5513a434f7eb-kube-api-access-g47wb\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546566 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546624 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546677 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546710 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.546753 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.547478 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-ca" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.561815 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648168 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648268 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648327 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648408 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648477 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648538 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g47wb\" (UniqueName: \"kubernetes.io/projected/9614d2ac-beb6-461b-99cf-5513a434f7eb-kube-api-access-g47wb\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648574 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648646 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648735 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648806 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648884 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.648969 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.649111 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.649173 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.649233 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.649522 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.649583 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.649861 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.650001 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.650079 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.650453 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.658910 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.669147 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.674960 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g47wb\" (UniqueName: \"kubernetes.io/projected/9614d2ac-beb6-461b-99cf-5513a434f7eb-kube-api-access-g47wb\") pod \"service-telemetry-operator-4-build\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:51 crc kubenswrapper[4690]: I0217 00:16:51.864990 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:52 crc kubenswrapper[4690]: I0217 00:16:52.409469 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 17 00:16:53 crc kubenswrapper[4690]: I0217 00:16:53.142780 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"9614d2ac-beb6-461b-99cf-5513a434f7eb","Type":"ContainerStarted","Data":"9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f"} Feb 17 00:16:53 crc kubenswrapper[4690]: I0217 00:16:53.143077 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"9614d2ac-beb6-461b-99cf-5513a434f7eb","Type":"ContainerStarted","Data":"88824a4639e997d3542ca7627808a98c6f2c59b32415e25b08ce5620f1ee3f71"} Feb 17 00:16:53 crc kubenswrapper[4690]: E0217 00:16:53.227503 4690 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=6832776695153960605, SKID=, AKID=F6:B0:84:3E:CE:5D:F7:B6:01:CF:ED:4A:AE:12:EA:43:DB:35:C5:B4 failed: x509: certificate signed by unknown authority" Feb 17 00:16:54 crc kubenswrapper[4690]: I0217 00:16:54.262611 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.158863 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-4-build" podUID="9614d2ac-beb6-461b-99cf-5513a434f7eb" containerName="git-clone" containerID="cri-o://9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f" gracePeriod=30 Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.568471 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_9614d2ac-beb6-461b-99cf-5513a434f7eb/git-clone/0.log" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.568863 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.608393 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-ca-bundles\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.608463 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-node-pullsecrets\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.608520 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-blob-cache\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.608587 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609037 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609129 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-root\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609334 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609566 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609640 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildcachedir\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609705 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-run\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609737 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-pull\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609779 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildworkdir\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609826 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-push\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609870 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g47wb\" (UniqueName: \"kubernetes.io/projected/9614d2ac-beb6-461b-99cf-5513a434f7eb-kube-api-access-g47wb\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609914 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-system-configs\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.609983 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-proxy-ca-bundles\") pod \"9614d2ac-beb6-461b-99cf-5513a434f7eb\" (UID: \"9614d2ac-beb6-461b-99cf-5513a434f7eb\") " Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.610312 4690 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.610337 4690 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.610355 4690 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.610397 4690 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.610827 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.611119 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.611476 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.612722 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.615061 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.619084 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-push" (OuterVolumeSpecName: "builder-dockercfg-nkjdw-push") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "builder-dockercfg-nkjdw-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.619161 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-pull" (OuterVolumeSpecName: "builder-dockercfg-nkjdw-pull") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "builder-dockercfg-nkjdw-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.622610 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9614d2ac-beb6-461b-99cf-5513a434f7eb-kube-api-access-g47wb" (OuterVolumeSpecName: "kube-api-access-g47wb") pod "9614d2ac-beb6-461b-99cf-5513a434f7eb" (UID: "9614d2ac-beb6-461b-99cf-5513a434f7eb"). InnerVolumeSpecName "kube-api-access-g47wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.714978 4690 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.715024 4690 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.715036 4690 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-pull\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.715046 4690 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9614d2ac-beb6-461b-99cf-5513a434f7eb-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.715055 4690 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/9614d2ac-beb6-461b-99cf-5513a434f7eb-builder-dockercfg-nkjdw-push\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.715065 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g47wb\" (UniqueName: \"kubernetes.io/projected/9614d2ac-beb6-461b-99cf-5513a434f7eb-kube-api-access-g47wb\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.715073 4690 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:55 crc kubenswrapper[4690]: I0217 00:16:55.715082 4690 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9614d2ac-beb6-461b-99cf-5513a434f7eb-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.171201 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_9614d2ac-beb6-461b-99cf-5513a434f7eb/git-clone/0.log" Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.171314 4690 generic.go:334] "Generic (PLEG): container finished" podID="9614d2ac-beb6-461b-99cf-5513a434f7eb" containerID="9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f" exitCode=1 Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.171384 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"9614d2ac-beb6-461b-99cf-5513a434f7eb","Type":"ContainerDied","Data":"9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f"} Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.171452 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"9614d2ac-beb6-461b-99cf-5513a434f7eb","Type":"ContainerDied","Data":"88824a4639e997d3542ca7627808a98c6f2c59b32415e25b08ce5620f1ee3f71"} Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.171482 4690 scope.go:117] "RemoveContainer" containerID="9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f" Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.171545 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.202582 4690 scope.go:117] "RemoveContainer" containerID="9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f" Feb 17 00:16:56 crc kubenswrapper[4690]: E0217 00:16:56.204837 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f\": container with ID starting with 9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f not found: ID does not exist" containerID="9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f" Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.204921 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f"} err="failed to get container status \"9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f\": rpc error: code = NotFound desc = could not find container \"9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f\": container with ID starting with 9741c5ca517605c3e0dfea7880c5c27456df918a3885908e971e77a46ce89b3f not found: ID does not exist" Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.230617 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.240670 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.325018 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.325111 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:16:56 crc kubenswrapper[4690]: I0217 00:16:56.989010 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9614d2ac-beb6-461b-99cf-5513a434f7eb" path="/var/lib/kubelet/pods/9614d2ac-beb6-461b-99cf-5513a434f7eb/volumes" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.727413 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 17 00:17:05 crc kubenswrapper[4690]: E0217 00:17:05.728216 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9614d2ac-beb6-461b-99cf-5513a434f7eb" containerName="git-clone" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.728233 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9614d2ac-beb6-461b-99cf-5513a434f7eb" containerName="git-clone" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.728402 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="9614d2ac-beb6-461b-99cf-5513a434f7eb" containerName="git-clone" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.729628 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.733231 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-sys-config" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.733495 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-ca" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.733540 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-global-ca" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.735379 4690 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-nkjdw" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.749159 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863346 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863426 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863500 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863566 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863622 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863673 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nxlz\" (UniqueName: \"kubernetes.io/projected/a7beafad-91dc-4d09-b04f-212a0702d83c-kube-api-access-5nxlz\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863717 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863746 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863792 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863822 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863860 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.863916 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.964949 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965027 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965068 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965106 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965160 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965202 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nxlz\" (UniqueName: \"kubernetes.io/projected/a7beafad-91dc-4d09-b04f-212a0702d83c-kube-api-access-5nxlz\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965242 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965273 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965322 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965351 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965438 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965499 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965713 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.965894 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.966108 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.966440 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.966616 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.966651 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.967154 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.967264 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.968408 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.974772 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.974992 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-push\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:05 crc kubenswrapper[4690]: I0217 00:17:05.994762 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nxlz\" (UniqueName: \"kubernetes.io/projected/a7beafad-91dc-4d09-b04f-212a0702d83c-kube-api-access-5nxlz\") pod \"service-telemetry-operator-5-build\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:06 crc kubenswrapper[4690]: I0217 00:17:06.056455 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:06 crc kubenswrapper[4690]: I0217 00:17:06.559394 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 17 00:17:07 crc kubenswrapper[4690]: I0217 00:17:07.271116 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"a7beafad-91dc-4d09-b04f-212a0702d83c","Type":"ContainerStarted","Data":"b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b"} Feb 17 00:17:07 crc kubenswrapper[4690]: I0217 00:17:07.271661 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"a7beafad-91dc-4d09-b04f-212a0702d83c","Type":"ContainerStarted","Data":"0d456239ba7e24b364faf46539070b66f8162f485369c2ea52e682ef97a8883b"} Feb 17 00:17:07 crc kubenswrapper[4690]: E0217 00:17:07.352500 4690 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=6832776695153960605, SKID=, AKID=F6:B0:84:3E:CE:5D:F7:B6:01:CF:ED:4A:AE:12:EA:43:DB:35:C5:B4 failed: x509: certificate signed by unknown authority" Feb 17 00:17:08 crc kubenswrapper[4690]: I0217 00:17:08.385521 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.284935 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-5-build" podUID="a7beafad-91dc-4d09-b04f-212a0702d83c" containerName="git-clone" containerID="cri-o://b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b" gracePeriod=30 Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.670347 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-5-build_a7beafad-91dc-4d09-b04f-212a0702d83c/git-clone/0.log" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.670612 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821202 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-push\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821565 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-node-pullsecrets\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821617 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-buildcachedir\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821655 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-ca-bundles\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821690 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nxlz\" (UniqueName: \"kubernetes.io/projected/a7beafad-91dc-4d09-b04f-212a0702d83c-kube-api-access-5nxlz\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821719 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-build-blob-cache\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821751 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-root\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821780 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-proxy-ca-bundles\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821810 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-run\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821834 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-pull\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821858 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-system-configs\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.821888 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-buildworkdir\") pod \"a7beafad-91dc-4d09-b04f-212a0702d83c\" (UID: \"a7beafad-91dc-4d09-b04f-212a0702d83c\") " Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.822248 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.822262 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.822637 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.822786 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.823011 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.823176 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.823535 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.823836 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.823981 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.828693 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7beafad-91dc-4d09-b04f-212a0702d83c-kube-api-access-5nxlz" (OuterVolumeSpecName: "kube-api-access-5nxlz") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "kube-api-access-5nxlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.829548 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-push" (OuterVolumeSpecName: "builder-dockercfg-nkjdw-push") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "builder-dockercfg-nkjdw-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.830151 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-pull" (OuterVolumeSpecName: "builder-dockercfg-nkjdw-pull") pod "a7beafad-91dc-4d09-b04f-212a0702d83c" (UID: "a7beafad-91dc-4d09-b04f-212a0702d83c"). InnerVolumeSpecName "builder-dockercfg-nkjdw-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.922917 4690 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.923138 4690 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.923214 4690 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.923305 4690 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-nkjdw-pull\" (UniqueName: \"kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-pull\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.923504 4690 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.923582 4690 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.923653 4690 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-nkjdw-push\" (UniqueName: \"kubernetes.io/secret/a7beafad-91dc-4d09-b04f-212a0702d83c-builder-dockercfg-nkjdw-push\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.923724 4690 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.923795 4690 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a7beafad-91dc-4d09-b04f-212a0702d83c-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.923870 4690 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a7beafad-91dc-4d09-b04f-212a0702d83c-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.923940 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nxlz\" (UniqueName: \"kubernetes.io/projected/a7beafad-91dc-4d09-b04f-212a0702d83c-kube-api-access-5nxlz\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:09 crc kubenswrapper[4690]: I0217 00:17:09.924016 4690 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a7beafad-91dc-4d09-b04f-212a0702d83c-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.297262 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-5-build_a7beafad-91dc-4d09-b04f-212a0702d83c/git-clone/0.log" Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.297793 4690 generic.go:334] "Generic (PLEG): container finished" podID="a7beafad-91dc-4d09-b04f-212a0702d83c" containerID="b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b" exitCode=1 Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.297840 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"a7beafad-91dc-4d09-b04f-212a0702d83c","Type":"ContainerDied","Data":"b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b"} Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.297878 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"a7beafad-91dc-4d09-b04f-212a0702d83c","Type":"ContainerDied","Data":"0d456239ba7e24b364faf46539070b66f8162f485369c2ea52e682ef97a8883b"} Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.297907 4690 scope.go:117] "RemoveContainer" containerID="b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b" Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.297963 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.329585 4690 scope.go:117] "RemoveContainer" containerID="b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b" Feb 17 00:17:10 crc kubenswrapper[4690]: E0217 00:17:10.330070 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b\": container with ID starting with b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b not found: ID does not exist" containerID="b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b" Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.330123 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b"} err="failed to get container status \"b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b\": rpc error: code = NotFound desc = could not find container \"b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b\": container with ID starting with b7e865ad5c1504a75f26d843983960c598586a9d46f6be594325ebbb8fa8712b not found: ID does not exist" Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.358615 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.366285 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 17 00:17:10 crc kubenswrapper[4690]: I0217 00:17:10.987334 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7beafad-91dc-4d09-b04f-212a0702d83c" path="/var/lib/kubelet/pods/a7beafad-91dc-4d09-b04f-212a0702d83c/volumes" Feb 17 00:17:26 crc kubenswrapper[4690]: I0217 00:17:26.325687 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:17:26 crc kubenswrapper[4690]: I0217 00:17:26.326727 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:17:46 crc kubenswrapper[4690]: I0217 00:17:46.864847 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h8qgc/must-gather-mdhm6"] Feb 17 00:17:46 crc kubenswrapper[4690]: E0217 00:17:46.865965 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7beafad-91dc-4d09-b04f-212a0702d83c" containerName="git-clone" Feb 17 00:17:46 crc kubenswrapper[4690]: I0217 00:17:46.865987 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7beafad-91dc-4d09-b04f-212a0702d83c" containerName="git-clone" Feb 17 00:17:46 crc kubenswrapper[4690]: I0217 00:17:46.866149 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7beafad-91dc-4d09-b04f-212a0702d83c" containerName="git-clone" Feb 17 00:17:46 crc kubenswrapper[4690]: I0217 00:17:46.867056 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h8qgc/must-gather-mdhm6" Feb 17 00:17:46 crc kubenswrapper[4690]: I0217 00:17:46.870481 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-h8qgc"/"default-dockercfg-jngtb" Feb 17 00:17:46 crc kubenswrapper[4690]: I0217 00:17:46.870804 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h8qgc"/"kube-root-ca.crt" Feb 17 00:17:46 crc kubenswrapper[4690]: I0217 00:17:46.871049 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h8qgc"/"openshift-service-ca.crt" Feb 17 00:17:46 crc kubenswrapper[4690]: I0217 00:17:46.929672 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h8qgc/must-gather-mdhm6"] Feb 17 00:17:46 crc kubenswrapper[4690]: I0217 00:17:46.996930 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fczm\" (UniqueName: \"kubernetes.io/projected/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-kube-api-access-6fczm\") pod \"must-gather-mdhm6\" (UID: \"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2\") " pod="openshift-must-gather-h8qgc/must-gather-mdhm6" Feb 17 00:17:46 crc kubenswrapper[4690]: I0217 00:17:46.996993 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-must-gather-output\") pod \"must-gather-mdhm6\" (UID: \"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2\") " pod="openshift-must-gather-h8qgc/must-gather-mdhm6" Feb 17 00:17:47 crc kubenswrapper[4690]: I0217 00:17:47.098525 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fczm\" (UniqueName: \"kubernetes.io/projected/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-kube-api-access-6fczm\") pod \"must-gather-mdhm6\" (UID: \"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2\") " pod="openshift-must-gather-h8qgc/must-gather-mdhm6" Feb 17 00:17:47 crc kubenswrapper[4690]: I0217 00:17:47.098579 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-must-gather-output\") pod \"must-gather-mdhm6\" (UID: \"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2\") " pod="openshift-must-gather-h8qgc/must-gather-mdhm6" Feb 17 00:17:47 crc kubenswrapper[4690]: I0217 00:17:47.099568 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-must-gather-output\") pod \"must-gather-mdhm6\" (UID: \"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2\") " pod="openshift-must-gather-h8qgc/must-gather-mdhm6" Feb 17 00:17:47 crc kubenswrapper[4690]: I0217 00:17:47.117088 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fczm\" (UniqueName: \"kubernetes.io/projected/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-kube-api-access-6fczm\") pod \"must-gather-mdhm6\" (UID: \"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2\") " pod="openshift-must-gather-h8qgc/must-gather-mdhm6" Feb 17 00:17:47 crc kubenswrapper[4690]: I0217 00:17:47.200315 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h8qgc/must-gather-mdhm6" Feb 17 00:17:47 crc kubenswrapper[4690]: I0217 00:17:47.399009 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h8qgc/must-gather-mdhm6"] Feb 17 00:17:47 crc kubenswrapper[4690]: W0217 00:17:47.419615 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3da9581_64f2_43b1_94c6_eb0ec1ddfce2.slice/crio-eedc95b49c1dc56ea6e5a4d16b808c6bf83d93106833e94e8ef9563f0ddf3c6f WatchSource:0}: Error finding container eedc95b49c1dc56ea6e5a4d16b808c6bf83d93106833e94e8ef9563f0ddf3c6f: Status 404 returned error can't find the container with id eedc95b49c1dc56ea6e5a4d16b808c6bf83d93106833e94e8ef9563f0ddf3c6f Feb 17 00:17:47 crc kubenswrapper[4690]: I0217 00:17:47.590644 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h8qgc/must-gather-mdhm6" event={"ID":"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2","Type":"ContainerStarted","Data":"eedc95b49c1dc56ea6e5a4d16b808c6bf83d93106833e94e8ef9563f0ddf3c6f"} Feb 17 00:17:54 crc kubenswrapper[4690]: I0217 00:17:54.665636 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h8qgc/must-gather-mdhm6" event={"ID":"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2","Type":"ContainerStarted","Data":"59269f8324331b079c50119c86d295c2325ac57389d249ed6d3b90e891fe2148"} Feb 17 00:17:54 crc kubenswrapper[4690]: I0217 00:17:54.666442 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h8qgc/must-gather-mdhm6" event={"ID":"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2","Type":"ContainerStarted","Data":"851209d69a84cb42f6ae354f26969d11fa653c9896264c2c14fbf03bbec05fe1"} Feb 17 00:17:54 crc kubenswrapper[4690]: I0217 00:17:54.687814 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h8qgc/must-gather-mdhm6" podStartSLOduration=2.366128253 podStartE2EDuration="8.6877909s" podCreationTimestamp="2026-02-17 00:17:46 +0000 UTC" firstStartedPulling="2026-02-17 00:17:47.421620908 +0000 UTC m=+741.157333669" lastFinishedPulling="2026-02-17 00:17:53.743283555 +0000 UTC m=+747.478996316" observedRunningTime="2026-02-17 00:17:54.684897089 +0000 UTC m=+748.420609850" watchObservedRunningTime="2026-02-17 00:17:54.6877909 +0000 UTC m=+748.423503661" Feb 17 00:17:56 crc kubenswrapper[4690]: I0217 00:17:56.326158 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:17:56 crc kubenswrapper[4690]: I0217 00:17:56.326548 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:17:56 crc kubenswrapper[4690]: I0217 00:17:56.326609 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:17:56 crc kubenswrapper[4690]: I0217 00:17:56.327416 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"957be8da43ec9afba6c8ee251740bb0bf80c9b7194d697b255095f03095c6cae"} pod="openshift-machine-config-operator/machine-config-daemon-lszwm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 00:17:56 crc kubenswrapper[4690]: I0217 00:17:56.327500 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" containerID="cri-o://957be8da43ec9afba6c8ee251740bb0bf80c9b7194d697b255095f03095c6cae" gracePeriod=600 Feb 17 00:17:56 crc kubenswrapper[4690]: I0217 00:17:56.679511 4690 generic.go:334] "Generic (PLEG): container finished" podID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerID="957be8da43ec9afba6c8ee251740bb0bf80c9b7194d697b255095f03095c6cae" exitCode=0 Feb 17 00:17:56 crc kubenswrapper[4690]: I0217 00:17:56.679562 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerDied","Data":"957be8da43ec9afba6c8ee251740bb0bf80c9b7194d697b255095f03095c6cae"} Feb 17 00:17:56 crc kubenswrapper[4690]: I0217 00:17:56.679855 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerStarted","Data":"89728ec74ac9b5c5516bae0da6b58cfb266ce8ebbcb85158ba3987c808341024"} Feb 17 00:17:56 crc kubenswrapper[4690]: I0217 00:17:56.679878 4690 scope.go:117] "RemoveContainer" containerID="2bfdb84a832fb3eb0bc54fa1a1fa0b2de03c4db1987487e9015bfa4b3b7cb1bb" Feb 17 00:17:56 crc kubenswrapper[4690]: E0217 00:17:56.866785 4690 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=6832776695153960605, SKID=, AKID=F6:B0:84:3E:CE:5D:F7:B6:01:CF:ED:4A:AE:12:EA:43:DB:35:C5:B4 failed: x509: certificate signed by unknown authority" Feb 17 00:18:02 crc kubenswrapper[4690]: I0217 00:18:02.088918 4690 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 17 00:18:38 crc kubenswrapper[4690]: I0217 00:18:38.877305 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-7rclh_51b6988f-45a7-4faa-9f25-0c03e4fe316b/control-plane-machine-set-operator/0.log" Feb 17 00:18:39 crc kubenswrapper[4690]: I0217 00:18:39.026281 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vgzbg_0936c3dd-a635-4721-8499-cd99755c8de8/kube-rbac-proxy/0.log" Feb 17 00:18:39 crc kubenswrapper[4690]: I0217 00:18:39.053237 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vgzbg_0936c3dd-a635-4721-8499-cd99755c8de8/machine-api-operator/0.log" Feb 17 00:18:52 crc kubenswrapper[4690]: I0217 00:18:52.035603 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-kpkt9_01b55f46-e76a-4a56-a49c-888209f5b626/cert-manager-controller/0.log" Feb 17 00:18:52 crc kubenswrapper[4690]: I0217 00:18:52.201138 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-wtxfw_7728f7bd-93da-484e-b9df-2f4834cbd19b/cert-manager-cainjector/0.log" Feb 17 00:18:52 crc kubenswrapper[4690]: I0217 00:18:52.214669 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-58lzf_e9ca0053-6e23-41ba-b825-80acdae13e89/cert-manager-webhook/0.log" Feb 17 00:19:07 crc kubenswrapper[4690]: I0217 00:19:07.321814 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-hjdqb_80f87711-62d4-488b-8a93-3dc6d93e3cc4/prometheus-operator/0.log" Feb 17 00:19:07 crc kubenswrapper[4690]: I0217 00:19:07.494052 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg_f4c6248a-697b-47eb-b7af-603a2bdf7ec0/prometheus-operator-admission-webhook/0.log" Feb 17 00:19:07 crc kubenswrapper[4690]: I0217 00:19:07.509812 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj_dab3686a-8cc4-4105-8460-a534c5d7558d/prometheus-operator-admission-webhook/0.log" Feb 17 00:19:07 crc kubenswrapper[4690]: I0217 00:19:07.652723 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-7mmxh_bae5780c-307b-4f82-8759-bde07af42405/operator/0.log" Feb 17 00:19:07 crc kubenswrapper[4690]: I0217 00:19:07.683810 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-dj5w8_d6e1fe6c-28db-455c-b460-827567af661d/perses-operator/0.log" Feb 17 00:19:22 crc kubenswrapper[4690]: I0217 00:19:22.545396 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8_6c7e787f-4fb9-40ca-8ea8-331388ab61b0/util/0.log" Feb 17 00:19:22 crc kubenswrapper[4690]: I0217 00:19:22.677907 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8_6c7e787f-4fb9-40ca-8ea8-331388ab61b0/pull/0.log" Feb 17 00:19:22 crc kubenswrapper[4690]: I0217 00:19:22.711514 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8_6c7e787f-4fb9-40ca-8ea8-331388ab61b0/util/0.log" Feb 17 00:19:22 crc kubenswrapper[4690]: I0217 00:19:22.712172 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8_6c7e787f-4fb9-40ca-8ea8-331388ab61b0/pull/0.log" Feb 17 00:19:22 crc kubenswrapper[4690]: I0217 00:19:22.848226 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8_6c7e787f-4fb9-40ca-8ea8-331388ab61b0/util/0.log" Feb 17 00:19:22 crc kubenswrapper[4690]: I0217 00:19:22.864695 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8_6c7e787f-4fb9-40ca-8ea8-331388ab61b0/pull/0.log" Feb 17 00:19:22 crc kubenswrapper[4690]: I0217 00:19:22.876553 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1msvs8_6c7e787f-4fb9-40ca-8ea8-331388ab61b0/extract/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.015974 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl_73d09007-760b-47ba-938a-db3ec30b065f/util/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.164870 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl_73d09007-760b-47ba-938a-db3ec30b065f/util/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.166395 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl_73d09007-760b-47ba-938a-db3ec30b065f/pull/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.185233 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl_73d09007-760b-47ba-938a-db3ec30b065f/pull/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.345420 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl_73d09007-760b-47ba-938a-db3ec30b065f/extract/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.347172 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl_73d09007-760b-47ba-938a-db3ec30b065f/util/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.372900 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fmmxkl_73d09007-760b-47ba-938a-db3ec30b065f/pull/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.516613 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c_930903ed-43cf-4146-bf5a-56f6f9d16434/util/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.660488 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c_930903ed-43cf-4146-bf5a-56f6f9d16434/util/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.689658 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c_930903ed-43cf-4146-bf5a-56f6f9d16434/pull/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.747005 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c_930903ed-43cf-4146-bf5a-56f6f9d16434/pull/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.858338 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c_930903ed-43cf-4146-bf5a-56f6f9d16434/util/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.903294 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c_930903ed-43cf-4146-bf5a-56f6f9d16434/pull/0.log" Feb 17 00:19:23 crc kubenswrapper[4690]: I0217 00:19:23.911876 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5b8l4c_930903ed-43cf-4146-bf5a-56f6f9d16434/extract/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.017209 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64_584ec9e5-3927-451e-bd2c-5c5014425c24/util/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.192244 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64_584ec9e5-3927-451e-bd2c-5c5014425c24/pull/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.192396 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64_584ec9e5-3927-451e-bd2c-5c5014425c24/pull/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.198331 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64_584ec9e5-3927-451e-bd2c-5c5014425c24/util/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.413564 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64_584ec9e5-3927-451e-bd2c-5c5014425c24/util/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.418768 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64_584ec9e5-3927-451e-bd2c-5c5014425c24/extract/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.434104 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08rpv64_584ec9e5-3927-451e-bd2c-5c5014425c24/pull/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.558442 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4glwj_c8b18214-ded8-4377-81de-0c21887f2576/extract-utilities/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.720841 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4glwj_c8b18214-ded8-4377-81de-0c21887f2576/extract-content/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.741269 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4glwj_c8b18214-ded8-4377-81de-0c21887f2576/extract-utilities/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.743271 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4glwj_c8b18214-ded8-4377-81de-0c21887f2576/extract-content/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.883122 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v4bzj"] Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.884831 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.893458 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4bzj"] Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.912145 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4glwj_c8b18214-ded8-4377-81de-0c21887f2576/extract-utilities/0.log" Feb 17 00:19:24 crc kubenswrapper[4690]: I0217 00:19:24.923037 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4glwj_c8b18214-ded8-4377-81de-0c21887f2576/extract-content/0.log" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.040473 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqqbn\" (UniqueName: \"kubernetes.io/projected/ff437d46-04db-43b1-9f22-0f20cf7b31fd-kube-api-access-lqqbn\") pod \"certified-operators-v4bzj\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.040579 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-catalog-content\") pod \"certified-operators-v4bzj\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.040615 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-utilities\") pod \"certified-operators-v4bzj\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.093010 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4glwj_c8b18214-ded8-4377-81de-0c21887f2576/registry-server/0.log" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.142263 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqqbn\" (UniqueName: \"kubernetes.io/projected/ff437d46-04db-43b1-9f22-0f20cf7b31fd-kube-api-access-lqqbn\") pod \"certified-operators-v4bzj\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.142312 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-catalog-content\") pod \"certified-operators-v4bzj\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.142345 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-utilities\") pod \"certified-operators-v4bzj\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.142797 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-utilities\") pod \"certified-operators-v4bzj\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.142853 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-catalog-content\") pod \"certified-operators-v4bzj\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.159493 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lqg9n_9debb4ff-2d87-49ea-a168-8b403ef878cf/extract-utilities/0.log" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.162186 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqqbn\" (UniqueName: \"kubernetes.io/projected/ff437d46-04db-43b1-9f22-0f20cf7b31fd-kube-api-access-lqqbn\") pod \"certified-operators-v4bzj\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.202429 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.423046 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lqg9n_9debb4ff-2d87-49ea-a168-8b403ef878cf/extract-content/0.log" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.483243 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4bzj"] Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.552175 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lqg9n_9debb4ff-2d87-49ea-a168-8b403ef878cf/extract-utilities/0.log" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.582397 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lqg9n_9debb4ff-2d87-49ea-a168-8b403ef878cf/extract-content/0.log" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.756409 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lqg9n_9debb4ff-2d87-49ea-a168-8b403ef878cf/extract-utilities/0.log" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.824001 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lqg9n_9debb4ff-2d87-49ea-a168-8b403ef878cf/extract-content/0.log" Feb 17 00:19:25 crc kubenswrapper[4690]: I0217 00:19:25.991064 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lqg9n_9debb4ff-2d87-49ea-a168-8b403ef878cf/registry-server/0.log" Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.045892 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wvvgs_4372c018-0aaf-459e-9e8f-6423d7ed5695/marketplace-operator/0.log" Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.125672 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qnd7w_2cb4a9ed-9d67-4635-9083-9c6e383f148f/extract-utilities/0.log" Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.215240 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qnd7w_2cb4a9ed-9d67-4635-9083-9c6e383f148f/extract-utilities/0.log" Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.247551 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qnd7w_2cb4a9ed-9d67-4635-9083-9c6e383f148f/extract-content/0.log" Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.258287 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qnd7w_2cb4a9ed-9d67-4635-9083-9c6e383f148f/extract-content/0.log" Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.318694 4690 generic.go:334] "Generic (PLEG): container finished" podID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerID="e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701" exitCode=0 Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.318736 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4bzj" event={"ID":"ff437d46-04db-43b1-9f22-0f20cf7b31fd","Type":"ContainerDied","Data":"e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701"} Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.318767 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4bzj" event={"ID":"ff437d46-04db-43b1-9f22-0f20cf7b31fd","Type":"ContainerStarted","Data":"d6042f2522c93f371fd3a8471ed138790f9ca31ce933c75c19318dffa81ce7ec"} Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.380351 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qnd7w_2cb4a9ed-9d67-4635-9083-9c6e383f148f/extract-content/0.log" Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.380374 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qnd7w_2cb4a9ed-9d67-4635-9083-9c6e383f148f/extract-utilities/0.log" Feb 17 00:19:26 crc kubenswrapper[4690]: I0217 00:19:26.600999 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qnd7w_2cb4a9ed-9d67-4635-9083-9c6e383f148f/registry-server/0.log" Feb 17 00:19:27 crc kubenswrapper[4690]: E0217 00:19:27.749078 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff437d46_04db_43b1_9f22_0f20cf7b31fd.slice/crio-29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c.scope\": RecentStats: unable to find data in memory cache]" Feb 17 00:19:28 crc kubenswrapper[4690]: I0217 00:19:28.334578 4690 generic.go:334] "Generic (PLEG): container finished" podID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerID="29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c" exitCode=0 Feb 17 00:19:28 crc kubenswrapper[4690]: I0217 00:19:28.334636 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4bzj" event={"ID":"ff437d46-04db-43b1-9f22-0f20cf7b31fd","Type":"ContainerDied","Data":"29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c"} Feb 17 00:19:29 crc kubenswrapper[4690]: I0217 00:19:29.344165 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4bzj" event={"ID":"ff437d46-04db-43b1-9f22-0f20cf7b31fd","Type":"ContainerStarted","Data":"ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d"} Feb 17 00:19:32 crc kubenswrapper[4690]: E0217 00:19:32.090971 4690 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.116s" Feb 17 00:19:35 crc kubenswrapper[4690]: I0217 00:19:35.202941 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:35 crc kubenswrapper[4690]: I0217 00:19:35.203300 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:35 crc kubenswrapper[4690]: I0217 00:19:35.272262 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:35 crc kubenswrapper[4690]: I0217 00:19:35.293021 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v4bzj" podStartSLOduration=8.740448678 podStartE2EDuration="11.292995474s" podCreationTimestamp="2026-02-17 00:19:24 +0000 UTC" firstStartedPulling="2026-02-17 00:19:26.32072018 +0000 UTC m=+840.056432931" lastFinishedPulling="2026-02-17 00:19:28.873266936 +0000 UTC m=+842.608979727" observedRunningTime="2026-02-17 00:19:29.366794323 +0000 UTC m=+843.102507094" watchObservedRunningTime="2026-02-17 00:19:35.292995474 +0000 UTC m=+849.028708225" Feb 17 00:19:36 crc kubenswrapper[4690]: I0217 00:19:36.190489 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:36 crc kubenswrapper[4690]: I0217 00:19:36.249041 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4bzj"] Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.144103 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v4bzj" podUID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerName="registry-server" containerID="cri-o://ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d" gracePeriod=2 Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.568954 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.676456 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-catalog-content\") pod \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.676516 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-utilities\") pod \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.676597 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqqbn\" (UniqueName: \"kubernetes.io/projected/ff437d46-04db-43b1-9f22-0f20cf7b31fd-kube-api-access-lqqbn\") pod \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\" (UID: \"ff437d46-04db-43b1-9f22-0f20cf7b31fd\") " Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.679244 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-utilities" (OuterVolumeSpecName: "utilities") pod "ff437d46-04db-43b1-9f22-0f20cf7b31fd" (UID: "ff437d46-04db-43b1-9f22-0f20cf7b31fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.697021 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff437d46-04db-43b1-9f22-0f20cf7b31fd-kube-api-access-lqqbn" (OuterVolumeSpecName: "kube-api-access-lqqbn") pod "ff437d46-04db-43b1-9f22-0f20cf7b31fd" (UID: "ff437d46-04db-43b1-9f22-0f20cf7b31fd"). InnerVolumeSpecName "kube-api-access-lqqbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.778203 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.778230 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqqbn\" (UniqueName: \"kubernetes.io/projected/ff437d46-04db-43b1-9f22-0f20cf7b31fd-kube-api-access-lqqbn\") on node \"crc\" DevicePath \"\"" Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.804877 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff437d46-04db-43b1-9f22-0f20cf7b31fd" (UID: "ff437d46-04db-43b1-9f22-0f20cf7b31fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:19:38 crc kubenswrapper[4690]: I0217 00:19:38.879744 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff437d46-04db-43b1-9f22-0f20cf7b31fd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.151089 4690 generic.go:334] "Generic (PLEG): container finished" podID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerID="ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d" exitCode=0 Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.151131 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4bzj" Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.151151 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4bzj" event={"ID":"ff437d46-04db-43b1-9f22-0f20cf7b31fd","Type":"ContainerDied","Data":"ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d"} Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.151295 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4bzj" event={"ID":"ff437d46-04db-43b1-9f22-0f20cf7b31fd","Type":"ContainerDied","Data":"d6042f2522c93f371fd3a8471ed138790f9ca31ce933c75c19318dffa81ce7ec"} Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.151313 4690 scope.go:117] "RemoveContainer" containerID="ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d" Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.173582 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4bzj"] Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.174853 4690 scope.go:117] "RemoveContainer" containerID="29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c" Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.179729 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v4bzj"] Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.190671 4690 scope.go:117] "RemoveContainer" containerID="e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701" Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.213336 4690 scope.go:117] "RemoveContainer" containerID="ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d" Feb 17 00:19:39 crc kubenswrapper[4690]: E0217 00:19:39.213774 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d\": container with ID starting with ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d not found: ID does not exist" containerID="ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d" Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.213804 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d"} err="failed to get container status \"ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d\": rpc error: code = NotFound desc = could not find container \"ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d\": container with ID starting with ceceb9944126b3b7e3da498eb993b29b9b360f95097652e2d9f39b9c569cbe3d not found: ID does not exist" Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.213833 4690 scope.go:117] "RemoveContainer" containerID="29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c" Feb 17 00:19:39 crc kubenswrapper[4690]: E0217 00:19:39.214066 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c\": container with ID starting with 29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c not found: ID does not exist" containerID="29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c" Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.214093 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c"} err="failed to get container status \"29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c\": rpc error: code = NotFound desc = could not find container \"29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c\": container with ID starting with 29ab5b9d2a3d9988227593c22da00707261ca95cc51ab3dc620df118d523259c not found: ID does not exist" Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.214128 4690 scope.go:117] "RemoveContainer" containerID="e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701" Feb 17 00:19:39 crc kubenswrapper[4690]: E0217 00:19:39.214486 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701\": container with ID starting with e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701 not found: ID does not exist" containerID="e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701" Feb 17 00:19:39 crc kubenswrapper[4690]: I0217 00:19:39.214507 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701"} err="failed to get container status \"e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701\": rpc error: code = NotFound desc = could not find container \"e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701\": container with ID starting with e598e86a6f184361a4dfc1293d456299d88273edf3ee53399c30d6acd9b0b701 not found: ID does not exist" Feb 17 00:19:40 crc kubenswrapper[4690]: I0217 00:19:40.826189 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-hjdqb_80f87711-62d4-488b-8a93-3dc6d93e3cc4/prometheus-operator/0.log" Feb 17 00:19:40 crc kubenswrapper[4690]: I0217 00:19:40.878166 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-b74d65c59-8vdjj_dab3686a-8cc4-4105-8460-a534c5d7558d/prometheus-operator-admission-webhook/0.log" Feb 17 00:19:40 crc kubenswrapper[4690]: I0217 00:19:40.881985 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-b74d65c59-wwfrg_f4c6248a-697b-47eb-b7af-603a2bdf7ec0/prometheus-operator-admission-webhook/0.log" Feb 17 00:19:40 crc kubenswrapper[4690]: I0217 00:19:40.983569 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" path="/var/lib/kubelet/pods/ff437d46-04db-43b1-9f22-0f20cf7b31fd/volumes" Feb 17 00:19:40 crc kubenswrapper[4690]: I0217 00:19:40.998455 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-7mmxh_bae5780c-307b-4f82-8759-bde07af42405/operator/0.log" Feb 17 00:19:41 crc kubenswrapper[4690]: I0217 00:19:41.022642 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-dj5w8_d6e1fe6c-28db-455c-b460-827567af661d/perses-operator/0.log" Feb 17 00:19:56 crc kubenswrapper[4690]: I0217 00:19:56.324954 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:19:56 crc kubenswrapper[4690]: I0217 00:19:56.325595 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.750626 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bpdk9"] Feb 17 00:20:09 crc kubenswrapper[4690]: E0217 00:20:09.751493 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerName="registry-server" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.751514 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerName="registry-server" Feb 17 00:20:09 crc kubenswrapper[4690]: E0217 00:20:09.751547 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerName="extract-utilities" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.751560 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerName="extract-utilities" Feb 17 00:20:09 crc kubenswrapper[4690]: E0217 00:20:09.751582 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerName="extract-content" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.751596 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerName="extract-content" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.751792 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff437d46-04db-43b1-9f22-0f20cf7b31fd" containerName="registry-server" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.753135 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.762741 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bpdk9"] Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.879137 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l7fl\" (UniqueName: \"kubernetes.io/projected/c07eefd1-80dc-4450-8ff5-d99a747ac8c5-kube-api-access-9l7fl\") pod \"community-operators-bpdk9\" (UID: \"c07eefd1-80dc-4450-8ff5-d99a747ac8c5\") " pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.879242 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c07eefd1-80dc-4450-8ff5-d99a747ac8c5-utilities\") pod \"community-operators-bpdk9\" (UID: \"c07eefd1-80dc-4450-8ff5-d99a747ac8c5\") " pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.879413 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c07eefd1-80dc-4450-8ff5-d99a747ac8c5-catalog-content\") pod \"community-operators-bpdk9\" (UID: \"c07eefd1-80dc-4450-8ff5-d99a747ac8c5\") " pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.980388 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c07eefd1-80dc-4450-8ff5-d99a747ac8c5-catalog-content\") pod \"community-operators-bpdk9\" (UID: \"c07eefd1-80dc-4450-8ff5-d99a747ac8c5\") " pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.980651 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l7fl\" (UniqueName: \"kubernetes.io/projected/c07eefd1-80dc-4450-8ff5-d99a747ac8c5-kube-api-access-9l7fl\") pod \"community-operators-bpdk9\" (UID: \"c07eefd1-80dc-4450-8ff5-d99a747ac8c5\") " pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.980754 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c07eefd1-80dc-4450-8ff5-d99a747ac8c5-utilities\") pod \"community-operators-bpdk9\" (UID: \"c07eefd1-80dc-4450-8ff5-d99a747ac8c5\") " pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.980883 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c07eefd1-80dc-4450-8ff5-d99a747ac8c5-catalog-content\") pod \"community-operators-bpdk9\" (UID: \"c07eefd1-80dc-4450-8ff5-d99a747ac8c5\") " pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:09 crc kubenswrapper[4690]: I0217 00:20:09.981421 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c07eefd1-80dc-4450-8ff5-d99a747ac8c5-utilities\") pod \"community-operators-bpdk9\" (UID: \"c07eefd1-80dc-4450-8ff5-d99a747ac8c5\") " pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:10 crc kubenswrapper[4690]: I0217 00:20:10.005096 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l7fl\" (UniqueName: \"kubernetes.io/projected/c07eefd1-80dc-4450-8ff5-d99a747ac8c5-kube-api-access-9l7fl\") pod \"community-operators-bpdk9\" (UID: \"c07eefd1-80dc-4450-8ff5-d99a747ac8c5\") " pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:10 crc kubenswrapper[4690]: I0217 00:20:10.076173 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:10 crc kubenswrapper[4690]: I0217 00:20:10.346934 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bpdk9"] Feb 17 00:20:10 crc kubenswrapper[4690]: I0217 00:20:10.391305 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bpdk9" event={"ID":"c07eefd1-80dc-4450-8ff5-d99a747ac8c5","Type":"ContainerStarted","Data":"1a85da7e1e7d784d904b3e53451fa55133274088f9486262c1cb389bec75c673"} Feb 17 00:20:11 crc kubenswrapper[4690]: I0217 00:20:11.399422 4690 generic.go:334] "Generic (PLEG): container finished" podID="c07eefd1-80dc-4450-8ff5-d99a747ac8c5" containerID="c664fde7b138cd075b633a8b9ad94d3939a060670218f12527c1be51999fe735" exitCode=0 Feb 17 00:20:11 crc kubenswrapper[4690]: I0217 00:20:11.399524 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bpdk9" event={"ID":"c07eefd1-80dc-4450-8ff5-d99a747ac8c5","Type":"ContainerDied","Data":"c664fde7b138cd075b633a8b9ad94d3939a060670218f12527c1be51999fe735"} Feb 17 00:20:11 crc kubenswrapper[4690]: I0217 00:20:11.401981 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 00:20:15 crc kubenswrapper[4690]: I0217 00:20:15.429628 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bpdk9" event={"ID":"c07eefd1-80dc-4450-8ff5-d99a747ac8c5","Type":"ContainerStarted","Data":"ed1c29c1c7f74e3d9760ad8c75e25db666876a941c57df163bf843ba8158ff59"} Feb 17 00:20:16 crc kubenswrapper[4690]: I0217 00:20:16.449488 4690 generic.go:334] "Generic (PLEG): container finished" podID="c07eefd1-80dc-4450-8ff5-d99a747ac8c5" containerID="ed1c29c1c7f74e3d9760ad8c75e25db666876a941c57df163bf843ba8158ff59" exitCode=0 Feb 17 00:20:16 crc kubenswrapper[4690]: I0217 00:20:16.449563 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bpdk9" event={"ID":"c07eefd1-80dc-4450-8ff5-d99a747ac8c5","Type":"ContainerDied","Data":"ed1c29c1c7f74e3d9760ad8c75e25db666876a941c57df163bf843ba8158ff59"} Feb 17 00:20:18 crc kubenswrapper[4690]: I0217 00:20:18.468075 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bpdk9" event={"ID":"c07eefd1-80dc-4450-8ff5-d99a747ac8c5","Type":"ContainerStarted","Data":"63e4a96dc0dceee6d3a18e3a4b47b85d1c1f1facfa637c6b92bda3e5ebdf3d2e"} Feb 17 00:20:18 crc kubenswrapper[4690]: I0217 00:20:18.496168 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bpdk9" podStartSLOduration=3.517035753 podStartE2EDuration="9.496129744s" podCreationTimestamp="2026-02-17 00:20:09 +0000 UTC" firstStartedPulling="2026-02-17 00:20:11.401710147 +0000 UTC m=+885.137422898" lastFinishedPulling="2026-02-17 00:20:17.380804148 +0000 UTC m=+891.116516889" observedRunningTime="2026-02-17 00:20:18.495119246 +0000 UTC m=+892.230832027" watchObservedRunningTime="2026-02-17 00:20:18.496129744 +0000 UTC m=+892.231842545" Feb 17 00:20:20 crc kubenswrapper[4690]: I0217 00:20:20.077077 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:20 crc kubenswrapper[4690]: I0217 00:20:20.077341 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:21 crc kubenswrapper[4690]: I0217 00:20:21.133526 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-bpdk9" podUID="c07eefd1-80dc-4450-8ff5-d99a747ac8c5" containerName="registry-server" probeResult="failure" output=< Feb 17 00:20:21 crc kubenswrapper[4690]: timeout: failed to connect service ":50051" within 1s Feb 17 00:20:21 crc kubenswrapper[4690]: > Feb 17 00:20:26 crc kubenswrapper[4690]: I0217 00:20:26.326115 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:20:26 crc kubenswrapper[4690]: I0217 00:20:26.326627 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:20:30 crc kubenswrapper[4690]: I0217 00:20:30.124307 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:30 crc kubenswrapper[4690]: I0217 00:20:30.184577 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bpdk9" Feb 17 00:20:30 crc kubenswrapper[4690]: I0217 00:20:30.266222 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bpdk9"] Feb 17 00:20:30 crc kubenswrapper[4690]: I0217 00:20:30.363216 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lqg9n"] Feb 17 00:20:30 crc kubenswrapper[4690]: I0217 00:20:30.363691 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lqg9n" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerName="registry-server" containerID="cri-o://2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd" gracePeriod=2 Feb 17 00:20:30 crc kubenswrapper[4690]: I0217 00:20:30.456082 4690 generic.go:334] "Generic (PLEG): container finished" podID="c3da9581-64f2-43b1-94c6-eb0ec1ddfce2" containerID="851209d69a84cb42f6ae354f26969d11fa653c9896264c2c14fbf03bbec05fe1" exitCode=0 Feb 17 00:20:30 crc kubenswrapper[4690]: I0217 00:20:30.456190 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h8qgc/must-gather-mdhm6" event={"ID":"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2","Type":"ContainerDied","Data":"851209d69a84cb42f6ae354f26969d11fa653c9896264c2c14fbf03bbec05fe1"} Feb 17 00:20:30 crc kubenswrapper[4690]: I0217 00:20:30.457267 4690 scope.go:117] "RemoveContainer" containerID="851209d69a84cb42f6ae354f26969d11fa653c9896264c2c14fbf03bbec05fe1" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.152874 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.276910 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-utilities\") pod \"9debb4ff-2d87-49ea-a168-8b403ef878cf\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.277057 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfkg5\" (UniqueName: \"kubernetes.io/projected/9debb4ff-2d87-49ea-a168-8b403ef878cf-kube-api-access-sfkg5\") pod \"9debb4ff-2d87-49ea-a168-8b403ef878cf\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.277131 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-catalog-content\") pod \"9debb4ff-2d87-49ea-a168-8b403ef878cf\" (UID: \"9debb4ff-2d87-49ea-a168-8b403ef878cf\") " Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.292535 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9debb4ff-2d87-49ea-a168-8b403ef878cf-kube-api-access-sfkg5" (OuterVolumeSpecName: "kube-api-access-sfkg5") pod "9debb4ff-2d87-49ea-a168-8b403ef878cf" (UID: "9debb4ff-2d87-49ea-a168-8b403ef878cf"). InnerVolumeSpecName "kube-api-access-sfkg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.299162 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-utilities" (OuterVolumeSpecName: "utilities") pod "9debb4ff-2d87-49ea-a168-8b403ef878cf" (UID: "9debb4ff-2d87-49ea-a168-8b403ef878cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.329445 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9debb4ff-2d87-49ea-a168-8b403ef878cf" (UID: "9debb4ff-2d87-49ea-a168-8b403ef878cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.378633 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfkg5\" (UniqueName: \"kubernetes.io/projected/9debb4ff-2d87-49ea-a168-8b403ef878cf-kube-api-access-sfkg5\") on node \"crc\" DevicePath \"\"" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.378663 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.378673 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9debb4ff-2d87-49ea-a168-8b403ef878cf-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.463122 4690 generic.go:334] "Generic (PLEG): container finished" podID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerID="2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd" exitCode=0 Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.463169 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lqg9n" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.463199 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lqg9n" event={"ID":"9debb4ff-2d87-49ea-a168-8b403ef878cf","Type":"ContainerDied","Data":"2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd"} Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.463247 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lqg9n" event={"ID":"9debb4ff-2d87-49ea-a168-8b403ef878cf","Type":"ContainerDied","Data":"05b831284693454541df36a8e58a28d8f0b1be6baaf046688faaa7fd1e63d82c"} Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.463266 4690 scope.go:117] "RemoveContainer" containerID="2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.482750 4690 scope.go:117] "RemoveContainer" containerID="780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.487488 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lqg9n"] Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.498501 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lqg9n"] Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.518639 4690 scope.go:117] "RemoveContainer" containerID="6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.540299 4690 scope.go:117] "RemoveContainer" containerID="2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd" Feb 17 00:20:31 crc kubenswrapper[4690]: E0217 00:20:31.540739 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd\": container with ID starting with 2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd not found: ID does not exist" containerID="2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.540783 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd"} err="failed to get container status \"2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd\": rpc error: code = NotFound desc = could not find container \"2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd\": container with ID starting with 2d0e04a8d1df83b57cc157e81d4b5cbc8926d1ef7bd3cdcfde16ef5e17af37bd not found: ID does not exist" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.540810 4690 scope.go:117] "RemoveContainer" containerID="780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51" Feb 17 00:20:31 crc kubenswrapper[4690]: E0217 00:20:31.541166 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51\": container with ID starting with 780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51 not found: ID does not exist" containerID="780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.541197 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51"} err="failed to get container status \"780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51\": rpc error: code = NotFound desc = could not find container \"780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51\": container with ID starting with 780192ae97eee8a1985bfdff4f1a41f1221ad6c0519433aa9a1874c6db49be51 not found: ID does not exist" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.541218 4690 scope.go:117] "RemoveContainer" containerID="6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6" Feb 17 00:20:31 crc kubenswrapper[4690]: E0217 00:20:31.541493 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6\": container with ID starting with 6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6 not found: ID does not exist" containerID="6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6" Feb 17 00:20:31 crc kubenswrapper[4690]: I0217 00:20:31.541522 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6"} err="failed to get container status \"6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6\": rpc error: code = NotFound desc = could not find container \"6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6\": container with ID starting with 6c8f9b380f7b248ee53d30a0ce9206122b7bfbe27529f4a5da01edb6bc5fa2b6 not found: ID does not exist" Feb 17 00:20:32 crc kubenswrapper[4690]: I0217 00:20:32.989468 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" path="/var/lib/kubelet/pods/9debb4ff-2d87-49ea-a168-8b403ef878cf/volumes" Feb 17 00:20:36 crc kubenswrapper[4690]: I0217 00:20:36.894704 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h8qgc_must-gather-mdhm6_c3da9581-64f2-43b1-94c6-eb0ec1ddfce2/gather/0.log" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.718689 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lrqq5"] Feb 17 00:20:41 crc kubenswrapper[4690]: E0217 00:20:41.719446 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerName="extract-utilities" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.719460 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerName="extract-utilities" Feb 17 00:20:41 crc kubenswrapper[4690]: E0217 00:20:41.719484 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerName="extract-content" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.719492 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerName="extract-content" Feb 17 00:20:41 crc kubenswrapper[4690]: E0217 00:20:41.719508 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerName="registry-server" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.719516 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerName="registry-server" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.719666 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="9debb4ff-2d87-49ea-a168-8b403ef878cf" containerName="registry-server" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.720599 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.741164 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrqq5"] Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.836712 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-catalog-content\") pod \"redhat-operators-lrqq5\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.836829 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-utilities\") pod \"redhat-operators-lrqq5\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.836868 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzb7l\" (UniqueName: \"kubernetes.io/projected/b62bf7d8-c35e-4e07-91df-9d531b356629-kube-api-access-fzb7l\") pod \"redhat-operators-lrqq5\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.937710 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-utilities\") pod \"redhat-operators-lrqq5\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.937774 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzb7l\" (UniqueName: \"kubernetes.io/projected/b62bf7d8-c35e-4e07-91df-9d531b356629-kube-api-access-fzb7l\") pod \"redhat-operators-lrqq5\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.937817 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-catalog-content\") pod \"redhat-operators-lrqq5\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.938308 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-catalog-content\") pod \"redhat-operators-lrqq5\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.938648 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-utilities\") pod \"redhat-operators-lrqq5\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:41 crc kubenswrapper[4690]: I0217 00:20:41.964160 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzb7l\" (UniqueName: \"kubernetes.io/projected/b62bf7d8-c35e-4e07-91df-9d531b356629-kube-api-access-fzb7l\") pod \"redhat-operators-lrqq5\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:42 crc kubenswrapper[4690]: I0217 00:20:42.039399 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:42 crc kubenswrapper[4690]: I0217 00:20:42.324348 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrqq5"] Feb 17 00:20:42 crc kubenswrapper[4690]: I0217 00:20:42.825817 4690 generic.go:334] "Generic (PLEG): container finished" podID="b62bf7d8-c35e-4e07-91df-9d531b356629" containerID="6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01" exitCode=0 Feb 17 00:20:42 crc kubenswrapper[4690]: I0217 00:20:42.826009 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqq5" event={"ID":"b62bf7d8-c35e-4e07-91df-9d531b356629","Type":"ContainerDied","Data":"6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01"} Feb 17 00:20:42 crc kubenswrapper[4690]: I0217 00:20:42.826142 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqq5" event={"ID":"b62bf7d8-c35e-4e07-91df-9d531b356629","Type":"ContainerStarted","Data":"a89f7a15ff2ae8c8ec5ed55e223913e553b5b16b7c2548701654f2262627d5e1"} Feb 17 00:20:43 crc kubenswrapper[4690]: I0217 00:20:43.840132 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqq5" event={"ID":"b62bf7d8-c35e-4e07-91df-9d531b356629","Type":"ContainerStarted","Data":"eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5"} Feb 17 00:20:44 crc kubenswrapper[4690]: I0217 00:20:44.625410 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h8qgc/must-gather-mdhm6"] Feb 17 00:20:44 crc kubenswrapper[4690]: I0217 00:20:44.625945 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-h8qgc/must-gather-mdhm6" podUID="c3da9581-64f2-43b1-94c6-eb0ec1ddfce2" containerName="copy" containerID="cri-o://59269f8324331b079c50119c86d295c2325ac57389d249ed6d3b90e891fe2148" gracePeriod=2 Feb 17 00:20:44 crc kubenswrapper[4690]: I0217 00:20:44.629229 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h8qgc/must-gather-mdhm6"] Feb 17 00:20:44 crc kubenswrapper[4690]: I0217 00:20:44.852564 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h8qgc_must-gather-mdhm6_c3da9581-64f2-43b1-94c6-eb0ec1ddfce2/copy/0.log" Feb 17 00:20:44 crc kubenswrapper[4690]: I0217 00:20:44.860849 4690 generic.go:334] "Generic (PLEG): container finished" podID="c3da9581-64f2-43b1-94c6-eb0ec1ddfce2" containerID="59269f8324331b079c50119c86d295c2325ac57389d249ed6d3b90e891fe2148" exitCode=143 Feb 17 00:20:44 crc kubenswrapper[4690]: I0217 00:20:44.866022 4690 generic.go:334] "Generic (PLEG): container finished" podID="b62bf7d8-c35e-4e07-91df-9d531b356629" containerID="eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5" exitCode=0 Feb 17 00:20:44 crc kubenswrapper[4690]: I0217 00:20:44.866091 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqq5" event={"ID":"b62bf7d8-c35e-4e07-91df-9d531b356629","Type":"ContainerDied","Data":"eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5"} Feb 17 00:20:44 crc kubenswrapper[4690]: I0217 00:20:44.982244 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h8qgc_must-gather-mdhm6_c3da9581-64f2-43b1-94c6-eb0ec1ddfce2/copy/0.log" Feb 17 00:20:44 crc kubenswrapper[4690]: I0217 00:20:44.984683 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h8qgc/must-gather-mdhm6" Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.085393 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fczm\" (UniqueName: \"kubernetes.io/projected/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-kube-api-access-6fczm\") pod \"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2\" (UID: \"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2\") " Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.085585 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-must-gather-output\") pod \"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2\" (UID: \"c3da9581-64f2-43b1-94c6-eb0ec1ddfce2\") " Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.094483 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-kube-api-access-6fczm" (OuterVolumeSpecName: "kube-api-access-6fczm") pod "c3da9581-64f2-43b1-94c6-eb0ec1ddfce2" (UID: "c3da9581-64f2-43b1-94c6-eb0ec1ddfce2"). InnerVolumeSpecName "kube-api-access-6fczm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.133516 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c3da9581-64f2-43b1-94c6-eb0ec1ddfce2" (UID: "c3da9581-64f2-43b1-94c6-eb0ec1ddfce2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.186613 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fczm\" (UniqueName: \"kubernetes.io/projected/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-kube-api-access-6fczm\") on node \"crc\" DevicePath \"\"" Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.186650 4690 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.874516 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h8qgc_must-gather-mdhm6_c3da9581-64f2-43b1-94c6-eb0ec1ddfce2/copy/0.log" Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.876586 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h8qgc/must-gather-mdhm6" Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.876621 4690 scope.go:117] "RemoveContainer" containerID="59269f8324331b079c50119c86d295c2325ac57389d249ed6d3b90e891fe2148" Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.885759 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqq5" event={"ID":"b62bf7d8-c35e-4e07-91df-9d531b356629","Type":"ContainerStarted","Data":"ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894"} Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.901835 4690 scope.go:117] "RemoveContainer" containerID="851209d69a84cb42f6ae354f26969d11fa653c9896264c2c14fbf03bbec05fe1" Feb 17 00:20:45 crc kubenswrapper[4690]: I0217 00:20:45.908301 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lrqq5" podStartSLOduration=2.443406811 podStartE2EDuration="4.908283822s" podCreationTimestamp="2026-02-17 00:20:41 +0000 UTC" firstStartedPulling="2026-02-17 00:20:42.827887528 +0000 UTC m=+916.563600289" lastFinishedPulling="2026-02-17 00:20:45.292764549 +0000 UTC m=+919.028477300" observedRunningTime="2026-02-17 00:20:45.904151497 +0000 UTC m=+919.639864288" watchObservedRunningTime="2026-02-17 00:20:45.908283822 +0000 UTC m=+919.643996573" Feb 17 00:20:46 crc kubenswrapper[4690]: I0217 00:20:46.988534 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3da9581-64f2-43b1-94c6-eb0ec1ddfce2" path="/var/lib/kubelet/pods/c3da9581-64f2-43b1-94c6-eb0ec1ddfce2/volumes" Feb 17 00:20:52 crc kubenswrapper[4690]: I0217 00:20:52.040236 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:52 crc kubenswrapper[4690]: I0217 00:20:52.040894 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:20:53 crc kubenswrapper[4690]: I0217 00:20:53.100286 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lrqq5" podUID="b62bf7d8-c35e-4e07-91df-9d531b356629" containerName="registry-server" probeResult="failure" output=< Feb 17 00:20:53 crc kubenswrapper[4690]: timeout: failed to connect service ":50051" within 1s Feb 17 00:20:53 crc kubenswrapper[4690]: > Feb 17 00:20:56 crc kubenswrapper[4690]: I0217 00:20:56.324958 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:20:56 crc kubenswrapper[4690]: I0217 00:20:56.325456 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:20:56 crc kubenswrapper[4690]: I0217 00:20:56.325501 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" Feb 17 00:20:56 crc kubenswrapper[4690]: I0217 00:20:56.325917 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89728ec74ac9b5c5516bae0da6b58cfb266ce8ebbcb85158ba3987c808341024"} pod="openshift-machine-config-operator/machine-config-daemon-lszwm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 00:20:56 crc kubenswrapper[4690]: I0217 00:20:56.325967 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" containerID="cri-o://89728ec74ac9b5c5516bae0da6b58cfb266ce8ebbcb85158ba3987c808341024" gracePeriod=600 Feb 17 00:20:56 crc kubenswrapper[4690]: I0217 00:20:56.970413 4690 generic.go:334] "Generic (PLEG): container finished" podID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerID="89728ec74ac9b5c5516bae0da6b58cfb266ce8ebbcb85158ba3987c808341024" exitCode=0 Feb 17 00:20:56 crc kubenswrapper[4690]: I0217 00:20:56.970479 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerDied","Data":"89728ec74ac9b5c5516bae0da6b58cfb266ce8ebbcb85158ba3987c808341024"} Feb 17 00:20:56 crc kubenswrapper[4690]: I0217 00:20:56.970525 4690 scope.go:117] "RemoveContainer" containerID="957be8da43ec9afba6c8ee251740bb0bf80c9b7194d697b255095f03095c6cae" Feb 17 00:20:57 crc kubenswrapper[4690]: I0217 00:20:57.979425 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" event={"ID":"8868c633-9a82-4998-b1b4-06c6d76396ec","Type":"ContainerStarted","Data":"fe9f068c9e4a09c37f39879bc44410e313add47b401478daef11afc6d30ea5ca"} Feb 17 00:21:02 crc kubenswrapper[4690]: I0217 00:21:02.090035 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:21:02 crc kubenswrapper[4690]: I0217 00:21:02.132969 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:21:02 crc kubenswrapper[4690]: I0217 00:21:02.319707 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrqq5"] Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.022888 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lrqq5" podUID="b62bf7d8-c35e-4e07-91df-9d531b356629" containerName="registry-server" containerID="cri-o://ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894" gracePeriod=2 Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.433603 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.525524 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-utilities\") pod \"b62bf7d8-c35e-4e07-91df-9d531b356629\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.525607 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzb7l\" (UniqueName: \"kubernetes.io/projected/b62bf7d8-c35e-4e07-91df-9d531b356629-kube-api-access-fzb7l\") pod \"b62bf7d8-c35e-4e07-91df-9d531b356629\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.525656 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-catalog-content\") pod \"b62bf7d8-c35e-4e07-91df-9d531b356629\" (UID: \"b62bf7d8-c35e-4e07-91df-9d531b356629\") " Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.527116 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-utilities" (OuterVolumeSpecName: "utilities") pod "b62bf7d8-c35e-4e07-91df-9d531b356629" (UID: "b62bf7d8-c35e-4e07-91df-9d531b356629"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.531298 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b62bf7d8-c35e-4e07-91df-9d531b356629-kube-api-access-fzb7l" (OuterVolumeSpecName: "kube-api-access-fzb7l") pod "b62bf7d8-c35e-4e07-91df-9d531b356629" (UID: "b62bf7d8-c35e-4e07-91df-9d531b356629"). InnerVolumeSpecName "kube-api-access-fzb7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.626923 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.627226 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzb7l\" (UniqueName: \"kubernetes.io/projected/b62bf7d8-c35e-4e07-91df-9d531b356629-kube-api-access-fzb7l\") on node \"crc\" DevicePath \"\"" Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.704323 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b62bf7d8-c35e-4e07-91df-9d531b356629" (UID: "b62bf7d8-c35e-4e07-91df-9d531b356629"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 00:21:04 crc kubenswrapper[4690]: I0217 00:21:04.728610 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b62bf7d8-c35e-4e07-91df-9d531b356629-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.031657 4690 generic.go:334] "Generic (PLEG): container finished" podID="b62bf7d8-c35e-4e07-91df-9d531b356629" containerID="ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894" exitCode=0 Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.031720 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqq5" event={"ID":"b62bf7d8-c35e-4e07-91df-9d531b356629","Type":"ContainerDied","Data":"ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894"} Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.031750 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqq5" event={"ID":"b62bf7d8-c35e-4e07-91df-9d531b356629","Type":"ContainerDied","Data":"a89f7a15ff2ae8c8ec5ed55e223913e553b5b16b7c2548701654f2262627d5e1"} Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.031769 4690 scope.go:117] "RemoveContainer" containerID="ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894" Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.032799 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrqq5" Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.056103 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrqq5"] Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.056154 4690 scope.go:117] "RemoveContainer" containerID="eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5" Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.061645 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lrqq5"] Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.074620 4690 scope.go:117] "RemoveContainer" containerID="6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01" Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.093378 4690 scope.go:117] "RemoveContainer" containerID="ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894" Feb 17 00:21:05 crc kubenswrapper[4690]: E0217 00:21:05.093751 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894\": container with ID starting with ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894 not found: ID does not exist" containerID="ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894" Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.093808 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894"} err="failed to get container status \"ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894\": rpc error: code = NotFound desc = could not find container \"ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894\": container with ID starting with ef2ee2b4dc561eeb1ee0f46c234bf043f242930e139fd1b9fb0699a156098894 not found: ID does not exist" Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.093844 4690 scope.go:117] "RemoveContainer" containerID="eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5" Feb 17 00:21:05 crc kubenswrapper[4690]: E0217 00:21:05.094147 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5\": container with ID starting with eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5 not found: ID does not exist" containerID="eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5" Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.094185 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5"} err="failed to get container status \"eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5\": rpc error: code = NotFound desc = could not find container \"eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5\": container with ID starting with eca75a7148106b95acc09139ca5d9d1d13085b1755a6db1f9a331ae9f902d8d5 not found: ID does not exist" Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.094210 4690 scope.go:117] "RemoveContainer" containerID="6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01" Feb 17 00:21:05 crc kubenswrapper[4690]: E0217 00:21:05.094451 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01\": container with ID starting with 6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01 not found: ID does not exist" containerID="6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01" Feb 17 00:21:05 crc kubenswrapper[4690]: I0217 00:21:05.094481 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01"} err="failed to get container status \"6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01\": rpc error: code = NotFound desc = could not find container \"6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01\": container with ID starting with 6487a4e85e14f35afe6e5aa02944bbc8a81d8befd25075956fd30ae574af2c01 not found: ID does not exist" Feb 17 00:21:06 crc kubenswrapper[4690]: I0217 00:21:06.987123 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b62bf7d8-c35e-4e07-91df-9d531b356629" path="/var/lib/kubelet/pods/b62bf7d8-c35e-4e07-91df-9d531b356629/volumes" Feb 17 00:23:26 crc kubenswrapper[4690]: I0217 00:23:26.325675 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:23:26 crc kubenswrapper[4690]: I0217 00:23:26.326283 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 00:23:56 crc kubenswrapper[4690]: I0217 00:23:56.325533 4690 patch_prober.go:28] interesting pod/machine-config-daemon-lszwm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 00:23:56 crc kubenswrapper[4690]: I0217 00:23:56.326529 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lszwm" podUID="8868c633-9a82-4998-b1b4-06c6d76396ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515144732456024460 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015144732457017376 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015144730002016501 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015144730002015451 5ustar corecore